[ 464.923766] env[61824]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61824) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 464.924123] env[61824]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61824) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 464.924246] env[61824]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61824) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 464.924583] env[61824]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 465.018366] env[61824]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61824) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 465.028523] env[61824]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61824) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 465.635447] env[61824]: INFO nova.virt.driver [None req-76ad281c-8279-422b-8f3e-9ff4f316849a None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 465.704996] env[61824]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 465.705245] env[61824]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 465.705360] env[61824]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61824) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 468.818364] env[61824]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-8aea6de4-87d3-470e-9c84-c65ca7c6d888 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.835138] env[61824]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61824) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 468.835354] env[61824]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-49b8dcd1-6008-4921-a1c7-bf36d9479531 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.872832] env[61824]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 0f7cd. [ 468.873041] env[61824]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.168s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 468.873614] env[61824]: INFO nova.virt.vmwareapi.driver [None req-76ad281c-8279-422b-8f3e-9ff4f316849a None None] VMware vCenter version: 7.0.3 [ 468.877198] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd90140-a900-45cc-baf1-6b5393ed23de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.896113] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c482680-e8e3-4780-a32c-1e09aaccc906 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.902293] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ddcfe6-d8f0-4241-b3e8-81cce8c86679 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.909083] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafa284e-83c7-44b3-a22a-b2564633c21b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.922354] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75224d48-ee81-4fbc-ad0c-771dfb803620 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.928309] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc239314-9df7-41b0-b542-fe15c3f886c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.958594] env[61824]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-e5e15ce6-a101-48c1-8e9e-f5ae5efaabb5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 468.963851] env[61824]: DEBUG nova.virt.vmwareapi.driver [None req-76ad281c-8279-422b-8f3e-9ff4f316849a None None] Extension org.openstack.compute already exists. {{(pid=61824) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 468.966527] env[61824]: INFO nova.compute.provider_config [None req-76ad281c-8279-422b-8f3e-9ff4f316849a None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 469.469875] env[61824]: DEBUG nova.context [None req-76ad281c-8279-422b-8f3e-9ff4f316849a None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),5fcfc2d1-7771-4994-aa51-ca68cdc6de9c(cell1) {{(pid=61824) load_cells /opt/stack/nova/nova/context.py:464}} [ 469.472042] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 469.472268] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 469.472920] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.473348] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Acquiring lock "5fcfc2d1-7771-4994-aa51-ca68cdc6de9c" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 469.473625] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Lock "5fcfc2d1-7771-4994-aa51-ca68cdc6de9c" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 469.474762] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Lock "5fcfc2d1-7771-4994-aa51-ca68cdc6de9c" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.494944] env[61824]: INFO dbcounter [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Registered counter for database nova_cell0 [ 469.503154] env[61824]: INFO dbcounter [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Registered counter for database nova_cell1 [ 469.506628] env[61824]: DEBUG oslo_db.sqlalchemy.engines [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61824) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 469.507326] env[61824]: DEBUG oslo_db.sqlalchemy.engines [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61824) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 469.511888] env[61824]: ERROR nova.db.main.api [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 469.511888] env[61824]: result = function(*args, **kwargs) [ 469.511888] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 469.511888] env[61824]: return func(*args, **kwargs) [ 469.511888] env[61824]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 469.511888] env[61824]: result = fn(*args, **kwargs) [ 469.511888] env[61824]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 469.511888] env[61824]: return f(*args, **kwargs) [ 469.511888] env[61824]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 469.511888] env[61824]: return db.service_get_minimum_version(context, binaries) [ 469.511888] env[61824]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 469.511888] env[61824]: _check_db_access() [ 469.511888] env[61824]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 469.511888] env[61824]: stacktrace = ''.join(traceback.format_stack()) [ 469.511888] env[61824]: [ 469.513170] env[61824]: ERROR nova.db.main.api [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 469.513170] env[61824]: result = function(*args, **kwargs) [ 469.513170] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 469.513170] env[61824]: return func(*args, **kwargs) [ 469.513170] env[61824]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 469.513170] env[61824]: result = fn(*args, **kwargs) [ 469.513170] env[61824]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 469.513170] env[61824]: return f(*args, **kwargs) [ 469.513170] env[61824]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 469.513170] env[61824]: return db.service_get_minimum_version(context, binaries) [ 469.513170] env[61824]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 469.513170] env[61824]: _check_db_access() [ 469.513170] env[61824]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 469.513170] env[61824]: stacktrace = ''.join(traceback.format_stack()) [ 469.513170] env[61824]: [ 469.513780] env[61824]: WARNING nova.objects.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 469.513959] env[61824]: WARNING nova.objects.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Failed to get minimum service version for cell 5fcfc2d1-7771-4994-aa51-ca68cdc6de9c [ 469.514445] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Acquiring lock "singleton_lock" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 469.514612] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Acquired lock "singleton_lock" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 469.514855] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Releasing lock "singleton_lock" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 469.515180] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Full set of CONF: {{(pid=61824) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 469.515333] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ******************************************************************************** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 469.515491] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Configuration options gathered from: {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 469.515634] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 469.515831] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 469.515960] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ================================================================================ {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 469.516183] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] allow_resize_to_same_host = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.516353] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] arq_binding_timeout = 300 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.516486] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] backdoor_port = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.516613] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] backdoor_socket = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.516775] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] block_device_allocate_retries = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.516933] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] block_device_allocate_retries_interval = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.517125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cert = self.pem {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.517293] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.517461] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute_monitors = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.517636] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] config_dir = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.517805] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] config_drive_format = iso9660 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.517936] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.518114] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] config_source = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.518283] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] console_host = devstack {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.518447] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] control_exchange = nova {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.518605] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cpu_allocation_ratio = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.518765] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] daemon = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.518932] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] debug = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.519101] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] default_access_ip_network_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.519310] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] default_availability_zone = nova {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.519428] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] default_ephemeral_format = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.519587] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] default_green_pool_size = 1000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.519821] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.519983] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] default_schedule_zone = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.520153] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] disk_allocation_ratio = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.520315] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] enable_new_services = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.520490] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] enabled_apis = ['osapi_compute'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.520658] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] enabled_ssl_apis = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.520820] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] flat_injected = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.520977] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] force_config_drive = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.521155] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] force_raw_images = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.521323] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] graceful_shutdown_timeout = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.521483] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] heal_instance_info_cache_interval = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.521694] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] host = cpu-1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.521870] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.522044] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.522209] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.522430] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.522590] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instance_build_timeout = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.522749] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instance_delete_interval = 300 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.522915] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instance_format = [instance: %(uuid)s] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.523090] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instance_name_template = instance-%08x {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.523255] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instance_usage_audit = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.523424] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instance_usage_audit_period = month {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.523590] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.523755] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.523920] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] internal_service_availability_zone = internal {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.524086] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] key = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.524248] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] live_migration_retry_count = 30 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.524444] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_color = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.524609] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_config_append = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.524776] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.524935] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_dir = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.525104] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.525235] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_options = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.525416] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_rotate_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.525598] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_rotate_interval_type = days {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.525763] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] log_rotation_type = none {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.525898] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.526037] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.526208] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.526372] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.526506] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.526669] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] long_rpc_timeout = 1800 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.526827] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] max_concurrent_builds = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.526986] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] max_concurrent_live_migrations = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.527155] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] max_concurrent_snapshots = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.527313] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] max_local_block_devices = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.527533] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] max_logfile_count = 30 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.527721] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] max_logfile_size_mb = 200 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.527885] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] maximum_instance_delete_attempts = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.528062] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] metadata_listen = 0.0.0.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.528233] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] metadata_listen_port = 8775 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.528401] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] metadata_workers = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.528566] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] migrate_max_retries = -1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.528732] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] mkisofs_cmd = genisoimage {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.528939] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.529085] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] my_ip = 10.180.1.21 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.529251] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] network_allocate_retries = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.529427] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.529594] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.529755] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] osapi_compute_listen_port = 8774 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.529920] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] osapi_compute_unique_server_name_scope = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.530097] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] osapi_compute_workers = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.530262] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] password_length = 12 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.530904] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] periodic_enable = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.530904] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] periodic_fuzzy_delay = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.530904] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] pointer_model = usbtablet {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.531046] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] preallocate_images = none {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.531160] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] publish_errors = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.531292] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] pybasedir = /opt/stack/nova {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.531449] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ram_allocation_ratio = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.531609] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] rate_limit_burst = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.531773] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] rate_limit_except_level = CRITICAL {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.531929] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] rate_limit_interval = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.532117] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] reboot_timeout = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.532258] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] reclaim_instance_interval = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.532404] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] record = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.532568] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] reimage_timeout_per_gb = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.532730] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] report_interval = 120 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.532888] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] rescue_timeout = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.533056] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] reserved_host_cpus = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.533216] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] reserved_host_disk_mb = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.533371] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] reserved_host_memory_mb = 512 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.533528] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] reserved_huge_pages = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.533684] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] resize_confirm_window = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.533838] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] resize_fs_using_block_device = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.533993] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] resume_guests_state_on_host_boot = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.534170] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.534344] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] rpc_response_timeout = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.534515] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] run_external_periodic_tasks = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.534686] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] running_deleted_instance_action = reap {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.534844] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.535007] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] running_deleted_instance_timeout = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.535168] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler_instance_sync_interval = 120 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.535339] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_down_time = 720 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.535523] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] servicegroup_driver = db {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.535679] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] shell_completion = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.535839] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] shelved_offload_time = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.536049] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] shelved_poll_interval = 3600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.536241] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] shutdown_timeout = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.536407] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] source_is_ipv6 = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.536569] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ssl_only = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.536815] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.536982] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] sync_power_state_interval = 600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.537156] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] sync_power_state_pool_size = 1000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.537324] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] syslog_log_facility = LOG_USER {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.537479] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] tempdir = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.537640] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] timeout_nbd = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.537803] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] transport_url = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.537963] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] update_resources_interval = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.538134] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] use_cow_images = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.538294] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] use_eventlog = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.538448] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] use_journal = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.538606] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] use_json = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.538759] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] use_rootwrap_daemon = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.538914] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] use_stderr = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.539081] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] use_syslog = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.539355] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vcpu_pin_set = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.539406] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plugging_is_fatal = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.539604] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plugging_timeout = 300 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.539791] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] virt_mkfs = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.539955] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] volume_usage_poll_interval = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.540131] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] watch_log_file = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.540301] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] web = /usr/share/spice-html5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 469.540488] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.540658] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.540821] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.540993] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_concurrency.disable_process_locking = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.541556] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.541753] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.541927] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.542117] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.542295] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.542464] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.542647] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.auth_strategy = keystone {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.542818] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.compute_link_prefix = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.542992] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.543179] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.dhcp_domain = novalocal {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.543349] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.enable_instance_password = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.543520] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.glance_link_prefix = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.543712] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.543951] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.544142] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.instance_list_per_project_cells = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.544313] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.list_records_by_skipping_down_cells = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.544508] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.local_metadata_per_cell = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.544685] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.max_limit = 1000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.544856] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.metadata_cache_expiration = 15 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.545043] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.neutron_default_tenant_id = default {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.545268] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.response_validation = warn {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.545460] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.use_neutron_default_nets = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.545651] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.545804] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.545975] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.546165] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.546342] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.vendordata_dynamic_targets = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.546508] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.vendordata_jsonfile_path = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.546689] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.546883] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.backend = dogpile.cache.memcached {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.547059] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.backend_argument = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.547235] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.config_prefix = cache.oslo {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.547409] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.dead_timeout = 60.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.547573] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.debug_cache_backend = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.547736] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.enable_retry_client = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.547899] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.enable_socket_keepalive = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.548081] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.enabled = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.548247] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.enforce_fips_mode = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.548412] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.expiration_time = 600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.548578] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.hashclient_retry_attempts = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.548742] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.548902] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_dead_retry = 300 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.549071] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_password = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.549235] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.549398] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.549560] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_pool_maxsize = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.549722] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.549885] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_sasl_enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.550072] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.550328] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.550406] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.memcache_username = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.550582] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.proxies = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.550746] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.redis_db = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.550905] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.redis_password = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.551086] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.551267] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.551435] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.redis_server = localhost:6379 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.551600] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.redis_socket_timeout = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.551756] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.redis_username = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.551916] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.retry_attempts = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.552093] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.retry_delay = 0.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.552260] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.socket_keepalive_count = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.552423] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.socket_keepalive_idle = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.552585] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.socket_keepalive_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.552742] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.tls_allowed_ciphers = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.552899] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.tls_cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.553162] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.tls_certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.553233] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.tls_enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.553377] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cache.tls_keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.553547] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.553719] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.auth_type = password {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.553880] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.554065] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.554229] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.554417] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.554592] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.cross_az_attach = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.554755] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.debug = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.554916] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.endpoint_template = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.555090] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.http_retries = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.555252] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.555411] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.555579] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.os_region_name = RegionOne {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.555740] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.555900] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cinder.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.556079] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.556242] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.cpu_dedicated_set = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.556399] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.cpu_shared_set = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.556564] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.image_type_exclude_list = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.556743] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.556918] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.557094] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.557261] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.557433] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.557599] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.resource_provider_association_refresh = 300 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.557761] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.557921] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.shutdown_retry_interval = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.558111] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.558292] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] conductor.workers = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.558470] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] console.allowed_origins = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.558632] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] console.ssl_ciphers = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.558799] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] console.ssl_minimum_version = default {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.558970] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] consoleauth.enforce_session_timeout = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.559148] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] consoleauth.token_ttl = 600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.559372] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.559465] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.559630] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.559788] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.connect_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.559948] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.connect_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.560116] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.endpoint_override = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.560277] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.560432] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.560593] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.max_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.560750] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.min_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.560906] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.region_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.561073] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.retriable_status_codes = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.561233] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.service_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.561402] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.service_type = accelerator {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.561565] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.561721] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.status_code_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.561878] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.status_code_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.562043] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.562225] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.562387] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] cyborg.version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.562566] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.backend = sqlalchemy {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.562735] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.connection = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.562897] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.connection_debug = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.563075] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.connection_parameters = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.563242] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.connection_recycle_time = 3600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.563404] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.connection_trace = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.563567] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.db_inc_retry_interval = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.563726] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.db_max_retries = 20 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.563884] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.db_max_retry_interval = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.564053] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.db_retry_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.564219] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.max_overflow = 50 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.564400] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.max_pool_size = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.564568] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.max_retries = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.564738] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.564897] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.mysql_wsrep_sync_wait = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.565064] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.pool_timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.565229] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.retry_interval = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.565385] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.slave_connection = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.565544] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.sqlite_synchronous = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.565703] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] database.use_db_reconnect = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.565878] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.backend = sqlalchemy {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.566055] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.connection = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.566222] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.connection_debug = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.566391] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.connection_parameters = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.566559] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.connection_recycle_time = 3600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.566738] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.connection_trace = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.566912] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.db_inc_retry_interval = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.567087] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.db_max_retries = 20 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.567252] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.db_max_retry_interval = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.567415] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.db_retry_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.567577] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.max_overflow = 50 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.567736] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.max_pool_size = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.567896] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.max_retries = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.568075] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.568237] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.568392] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.pool_timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.568555] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.retry_interval = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.568711] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.slave_connection = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.568869] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] api_database.sqlite_synchronous = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.569058] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] devices.enabled_mdev_types = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.569239] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.569410] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.569574] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ephemeral_storage_encryption.enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.569737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.569905] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.api_servers = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.570078] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.570244] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.570406] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.570564] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.connect_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.570722] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.connect_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.570900] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.debug = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.571089] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.default_trusted_certificate_ids = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.571264] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.enable_certificate_validation = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.571427] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.enable_rbd_download = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.571586] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.endpoint_override = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.571748] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.571909] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.572079] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.max_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.572240] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.min_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.572403] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.num_retries = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.572574] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.rbd_ceph_conf = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.572751] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.rbd_connect_timeout = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.572925] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.rbd_pool = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.573195] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.rbd_user = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.573404] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.region_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.573575] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.retriable_status_codes = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.573738] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.service_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.573910] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.service_type = image {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.574090] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.574259] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.status_code_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.574443] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.status_code_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.574640] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.574826] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.574991] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.verify_glance_signatures = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.575168] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] glance.version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.575359] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] guestfs.debug = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.575509] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] mks.enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.575895] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.576177] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] image_cache.manager_interval = 2400 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.576487] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] image_cache.precache_concurrency = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.576709] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] image_cache.remove_unused_base_images = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.576895] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.577080] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.577262] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] image_cache.subdirectory_name = _base {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.577442] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.api_max_retries = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.577614] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.api_retry_interval = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.577776] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.577942] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.auth_type = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.578116] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.578280] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.578446] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.578612] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.conductor_group = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.578773] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.connect_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.578933] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.connect_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.579104] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.endpoint_override = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.579272] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.579432] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.579593] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.max_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.579751] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.min_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.579917] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.peer_list = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.580105] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.region_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.580272] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.retriable_status_codes = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.580498] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.serial_console_state_timeout = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.580591] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.service_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.580761] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.service_type = baremetal {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.580923] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.shard = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.581098] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.581262] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.status_code_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.581421] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.status_code_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.581584] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.581764] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.581925] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ironic.version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.582122] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.582300] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] key_manager.fixed_key = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.582484] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.582649] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.barbican_api_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.582809] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.barbican_endpoint = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.582979] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.barbican_endpoint_type = public {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.583153] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.barbican_region_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.583314] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.583471] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.583692] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.583829] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.583988] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.584164] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.number_of_retries = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.584326] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.retry_delay = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.584524] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.send_service_user_token = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.584695] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.584878] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.585053] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.verify_ssl = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.585218] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican.verify_ssl_path = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.585386] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.585571] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.auth_type = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.585738] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.585893] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.586068] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.586235] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.586393] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.586560] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.586719] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] barbican_service_user.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.586888] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.approle_role_id = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.587058] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.approle_secret_id = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.587234] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.kv_mountpoint = secret {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.587396] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.kv_path = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.587563] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.kv_version = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.587722] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.namespace = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.587883] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.root_token_id = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.588063] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.ssl_ca_crt_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.588245] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.timeout = 60.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.588410] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.use_ssl = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.588584] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.588762] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.588925] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.auth_type = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.589096] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.589258] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.589436] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.589615] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.connect_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.589777] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.connect_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.589935] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.endpoint_override = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.590109] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.590288] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.590422] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.max_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.590580] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.min_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.590737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.region_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.590896] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.retriable_status_codes = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.591063] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.service_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.591254] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.service_type = identity {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.591425] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.591582] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.status_code_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.591740] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.status_code_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.591913] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.592129] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.592295] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] keystone.version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.592497] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.connection_uri = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.592662] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.cpu_mode = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.592829] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.593027] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.cpu_models = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.593200] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.cpu_power_governor_high = performance {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.593372] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.593550] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.cpu_power_management = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.593730] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.593896] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.device_detach_attempts = 8 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.594068] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.device_detach_timeout = 20 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.594238] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.disk_cachemodes = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.594427] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.disk_prefix = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.594603] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.enabled_perf_events = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.594767] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.file_backed_memory = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.594933] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.gid_maps = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.595103] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.hw_disk_discard = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.595265] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.hw_machine_type = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.595441] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.images_rbd_ceph_conf = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.595637] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.595805] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.595973] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.images_rbd_glance_store_name = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.596153] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.images_rbd_pool = rbd {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.596323] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.images_type = default {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.596479] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.images_volume_group = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.596646] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.inject_key = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.596800] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.inject_partition = -2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.596957] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.inject_password = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.597129] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.iscsi_iface = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.597291] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.iser_use_multipath = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.597452] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.597617] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.597779] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_downtime = 500 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.597938] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.598112] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.598275] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_inbound_addr = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.598438] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.598609] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.598772] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_scheme = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.598947] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_timeout_action = abort {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.599124] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_tunnelled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.599287] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_uri = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.599451] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.live_migration_with_native_tls = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.599612] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.max_queues = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.599775] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.600031] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.600202] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.nfs_mount_options = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.600508] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.600686] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.600851] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.601019] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.601190] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.601355] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.num_pcie_ports = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.601525] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.601694] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.pmem_namespaces = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.601857] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.quobyte_client_cfg = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.602156] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.602336] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.602503] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.602675] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.602839] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rbd_secret_uuid = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.603009] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rbd_user = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.603187] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.603374] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.603540] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rescue_image_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.603702] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rescue_kernel_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.603863] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rescue_ramdisk_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.604043] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.604208] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.rx_queue_size = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.604400] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.smbfs_mount_options = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.604683] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.604858] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.snapshot_compression = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.605031] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.snapshot_image_format = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.605254] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.605434] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.sparse_logical_volumes = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.605621] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.swtpm_enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.605795] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.swtpm_group = tss {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.605964] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.swtpm_user = tss {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.606148] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.sysinfo_serial = unique {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.606310] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.tb_cache_size = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.606470] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.tx_queue_size = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.606636] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.uid_maps = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.606800] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.use_virtio_for_bridges = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.606972] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.virt_type = kvm {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.607155] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.volume_clear = zero {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.607320] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.volume_clear_size = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.607484] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.volume_use_multipath = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.607644] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.vzstorage_cache_path = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.607812] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.607979] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.608155] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.608324] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.608594] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.608773] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.vzstorage_mount_user = stack {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.608941] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.609130] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.609306] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.auth_type = password {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.609468] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.609631] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.609794] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.609954] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.connect_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.610125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.connect_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.610296] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.default_floating_pool = public {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.610455] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.endpoint_override = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.610622] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.extension_sync_interval = 600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.610783] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.http_retries = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.611017] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.611106] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.611267] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.max_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.611436] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.611597] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.min_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.611766] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.ovs_bridge = br-int {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.611932] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.physnets = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.612110] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.region_name = RegionOne {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.612271] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.retriable_status_codes = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.612438] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.service_metadata_proxy = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.612596] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.service_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.612760] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.service_type = network {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.612918] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.613085] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.status_code_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.613245] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.status_code_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.613403] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.613582] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.613743] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] neutron.version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.613915] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] notifications.bdms_in_notifications = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.614103] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] notifications.default_level = INFO {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.614277] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] notifications.notification_format = unversioned {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.614482] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] notifications.notify_on_state_change = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.614667] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.614843] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] pci.alias = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.615023] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] pci.device_spec = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.615192] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] pci.report_in_placement = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.615375] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.615567] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.auth_type = password {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.615744] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.615905] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.616072] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.616236] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.616392] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.connect_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.616552] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.connect_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.616707] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.default_domain_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.616885] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.default_domain_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.617109] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.domain_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.617279] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.domain_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.617441] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.endpoint_override = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.617604] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.617759] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.617915] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.max_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.618081] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.min_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.618251] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.password = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.618408] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.project_domain_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.618575] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.project_domain_name = Default {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.618753] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.project_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.618927] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.project_name = service {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.619106] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.region_name = RegionOne {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.619269] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.retriable_status_codes = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.619428] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.service_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.619595] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.service_type = placement {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.619756] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.619933] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.status_code_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.620117] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.status_code_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.620281] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.system_scope = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.620438] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.620595] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.trust_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.620752] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.user_domain_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.620917] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.user_domain_name = Default {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.621085] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.user_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.621259] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.username = nova {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.621440] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.621603] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] placement.version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.621836] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.cores = 20 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.622041] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.count_usage_from_placement = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.622223] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.622429] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.injected_file_content_bytes = 10240 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.622606] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.injected_file_path_length = 255 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.622775] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.injected_files = 5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.622944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.instances = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.623128] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.key_pairs = 100 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.623299] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.metadata_items = 128 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.623466] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.ram = 51200 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.623631] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.recheck_quota = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.623799] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.server_group_members = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.623966] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] quota.server_groups = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.624163] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.624345] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.624520] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.image_metadata_prefilter = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.624689] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.624853] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.max_attempts = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.625024] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.max_placement_results = 1000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.625191] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.625376] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.625571] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.625751] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] scheduler.workers = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.625936] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.626100] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.626280] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.626448] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.626615] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.626781] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.626944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.627145] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.627318] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.host_subset_size = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.627487] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.627649] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.627811] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.627978] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.isolated_hosts = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.628155] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.isolated_images = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.628336] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.628522] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.628693] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.628857] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.pci_in_placement = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.629032] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.629202] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.629365] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.629527] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.629688] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.629852] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.630021] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.track_instance_changes = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.630202] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.630373] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] metrics.required = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.630539] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] metrics.weight_multiplier = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.630703] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.630868] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] metrics.weight_setting = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.631196] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.631390] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] serial_console.enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.631583] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] serial_console.port_range = 10000:20000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.631755] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.631923] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.632103] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] serial_console.serialproxy_port = 6083 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.632272] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.632443] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.auth_type = password {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.632602] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.632758] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.632917] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.633086] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.633244] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.633411] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.send_service_user_token = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.633574] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.633744] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] service_user.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.633910] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.agent_enabled = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.634083] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.634415] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.634620] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.634793] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.html5proxy_port = 6082 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.634956] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.image_compression = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.635127] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.jpeg_compression = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.635285] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.playback_compression = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.635456] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.require_secure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.635636] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.server_listen = 127.0.0.1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.635806] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.635965] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.streaming_mode = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.636167] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] spice.zlib_compression = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.636345] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] upgrade_levels.baseapi = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.636518] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] upgrade_levels.compute = auto {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.636679] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] upgrade_levels.conductor = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.636833] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] upgrade_levels.scheduler = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.636997] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.637174] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.637330] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.637485] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.637647] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.637803] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.637958] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.638130] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.638286] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vendordata_dynamic_auth.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.638458] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.api_retry_count = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.638617] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.ca_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.638784] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.638949] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.cluster_name = testcl1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.639145] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.connection_pool_size = 10 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.639316] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.console_delay_seconds = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.639482] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.datastore_regex = ^datastore.* {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.639687] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.639857] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.host_password = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.640029] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.host_port = 443 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.640200] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.host_username = administrator@vsphere.local {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.640367] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.insecure = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.640526] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.integration_bridge = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.640688] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.maximum_objects = 100 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.640846] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.pbm_default_policy = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.641021] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.pbm_enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.641760] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.pbm_wsdl_location = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.641760] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.641760] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.serial_port_proxy_uri = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.641760] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.serial_port_service_uri = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.641882] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.task_poll_interval = 0.5 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.641997] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.use_linked_clone = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.642199] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.vnc_keymap = en-us {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.642371] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.vnc_port = 5900 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.642535] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vmware.vnc_port_total = 10000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.642717] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.auth_schemes = ['none'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.642888] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.643192] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.643378] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.643551] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.novncproxy_port = 6080 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.643727] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.server_listen = 127.0.0.1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.643895] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.644064] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.vencrypt_ca_certs = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.644222] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.vencrypt_client_cert = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.644396] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vnc.vencrypt_client_key = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.644592] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.644758] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.disable_deep_image_inspection = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.644917] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.645094] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.645278] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.645453] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.disable_rootwrap = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.645647] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.enable_numa_live_migration = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.645811] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.645974] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.646151] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.646317] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.libvirt_disable_apic = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.646474] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.646638] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.646797] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.646959] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.647131] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.647294] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.647455] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.647615] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.647774] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.647939] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.648148] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.648359] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.client_socket_timeout = 900 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.648537] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.default_pool_size = 1000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.648707] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.keep_alive = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.648873] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.max_header_line = 16384 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.649044] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.649209] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.ssl_ca_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.649376] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.ssl_cert_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.649530] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.ssl_key_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.649692] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.tcp_keepidle = 600 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.649867] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.650042] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] zvm.ca_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.650203] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] zvm.cloud_connector_url = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.650488] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.650662] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] zvm.reachable_timeout = 300 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.650839] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.enforce_new_defaults = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.651300] env[61824]: WARNING oslo_config.cfg [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 469.651506] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.enforce_scope = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.651689] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.policy_default_rule = default {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.651875] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.652063] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.policy_file = policy.yaml {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.652248] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.652411] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.652573] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.652732] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.652893] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.653072] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.653250] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.653428] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.connection_string = messaging:// {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.653596] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.enabled = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.653766] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.es_doc_type = notification {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.653929] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.es_scroll_size = 10000 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.654107] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.es_scroll_time = 2m {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.654299] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.filter_error_trace = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.654497] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.hmac_keys = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.654674] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.sentinel_service_name = mymaster {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.654845] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.socket_timeout = 0.1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.655018] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.trace_requests = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.655185] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler.trace_sqlalchemy = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.655378] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler_jaeger.process_tags = {} {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.655547] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler_jaeger.service_name_prefix = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.655718] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] profiler_otlp.service_name_prefix = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.655889] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] remote_debug.host = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.656061] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] remote_debug.port = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.656246] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.656413] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.656626] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.656801] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.656967] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657145] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657309] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657473] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657637] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657807] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.657962] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658147] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658319] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658489] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658660] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658829] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.658991] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659178] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659343] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659525] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659707] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.659875] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660048] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660219] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660381] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660544] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660707] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.660867] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661043] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661216] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.ssl = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661386] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661556] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661717] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.661884] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662064] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662233] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662450] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662624] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_notifications.retry = -1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662808] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.662985] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663168] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.auth_section = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663332] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.auth_type = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663490] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.cafile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663648] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.certfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663809] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.collect_timing = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.663964] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.connect_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664134] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.connect_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664289] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.endpoint_id = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664481] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.endpoint_override = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664653] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.insecure = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664811] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.keyfile = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.664967] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.max_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665137] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.min_version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665312] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.region_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665506] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.retriable_status_codes = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665675] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.service_name = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665834] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.service_type = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.665994] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.split_loggers = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666167] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.status_code_retries = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666326] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.status_code_retry_delay = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666483] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.timeout = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666682] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.valid_interfaces = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666793] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_limit.version = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.666958] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_reports.file_event_handler = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667136] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667296] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] oslo_reports.log_dir = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667466] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667626] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.667784] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670500] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670500] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670500] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670500] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670500] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_ovs_privileged.group = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670500] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.670500] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] vif_plug_ovs_privileged.user = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.671944] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_ovs.isolate_vif = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673125] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673431] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673431] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673431] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] os_vif_ovs.per_port_bridge = False {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673431] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] privsep_osbrick.capabilities = [21] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673431] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] privsep_osbrick.group = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673431] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] privsep_osbrick.helper_command = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673431] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] privsep_osbrick.user = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] nova_sys_admin.group = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] nova_sys_admin.helper_command = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.673737] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.674042] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] nova_sys_admin.user = None {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 469.674042] env[61824]: DEBUG oslo_service.service [None req-a7e7ebce-670b-4043-84e3-b6eeaf8b8dd2 None None] ******************************************************************************** {{(pid=61824) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 469.674409] env[61824]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 470.179412] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Getting list of instances from cluster (obj){ [ 470.179412] env[61824]: value = "domain-c8" [ 470.179412] env[61824]: _type = "ClusterComputeResource" [ 470.179412] env[61824]: } {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 470.179865] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e5de0a-ee28-438c-8953-752e1621f5e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.189945] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Got total of 0 instances {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 470.190186] env[61824]: WARNING nova.virt.vmwareapi.driver [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 470.191084] env[61824]: INFO nova.virt.node [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Generated node identity 8d9ece0e-729c-4d69-b741-d89969200860 [ 470.191084] env[61824]: INFO nova.virt.node [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Wrote node identity 8d9ece0e-729c-4d69-b741-d89969200860 to /opt/stack/data/n-cpu-1/compute_id [ 470.693657] env[61824]: WARNING nova.compute.manager [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Compute nodes ['8d9ece0e-729c-4d69-b741-d89969200860'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 471.699024] env[61824]: INFO nova.compute.manager [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 472.705329] env[61824]: WARNING nova.compute.manager [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 472.705707] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.705871] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 472.706034] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 472.706197] env[61824]: DEBUG nova.compute.resource_tracker [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 472.707171] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bdd7db-679d-473a-863d-83acf9506c82 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.715785] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb984bd-751e-4df7-bdc1-d9c23384eebc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.730540] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac755c4d-05ae-4b28-a37d-328c48554e1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.737165] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79670a58-caaf-480f-a616-8753a03bc2ee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 472.766564] env[61824]: DEBUG nova.compute.resource_tracker [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181536MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 472.766698] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 472.766889] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 473.269248] env[61824]: WARNING nova.compute.resource_tracker [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] No compute node record for cpu-1:8d9ece0e-729c-4d69-b741-d89969200860: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 8d9ece0e-729c-4d69-b741-d89969200860 could not be found. [ 473.774277] env[61824]: INFO nova.compute.resource_tracker [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 8d9ece0e-729c-4d69-b741-d89969200860 [ 475.281490] env[61824]: DEBUG nova.compute.resource_tracker [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 475.281860] env[61824]: DEBUG nova.compute.resource_tracker [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 475.440711] env[61824]: INFO nova.scheduler.client.report [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] [req-6356916b-28e7-4caf-b3d8-192ada74afef] Created resource provider record via placement API for resource provider with UUID 8d9ece0e-729c-4d69-b741-d89969200860 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 475.455851] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d55946c-6d68-4de2-807c-07032a4f202b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.463813] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7036b9c8-a7ea-4ea1-9f80-8e54bd506df8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.493614] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed3f600-4d3a-4936-a9e3-689fa68662a7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.501038] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e8c4b83-9b3b-48d5-9333-80a39027151c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 475.513952] env[61824]: DEBUG nova.compute.provider_tree [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.050756] env[61824]: DEBUG nova.scheduler.client.report [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Updated inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 476.051021] env[61824]: DEBUG nova.compute.provider_tree [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Updating resource provider 8d9ece0e-729c-4d69-b741-d89969200860 generation from 0 to 1 during operation: update_inventory {{(pid=61824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.051162] env[61824]: DEBUG nova.compute.provider_tree [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.098297] env[61824]: DEBUG nova.compute.provider_tree [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Updating resource provider 8d9ece0e-729c-4d69-b741-d89969200860 generation from 1 to 2 during operation: update_traits {{(pid=61824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.602744] env[61824]: DEBUG nova.compute.resource_tracker [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 476.603110] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.836s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 476.603156] env[61824]: DEBUG nova.service [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Creating RPC server for service compute {{(pid=61824) start /opt/stack/nova/nova/service.py:186}} [ 476.617660] env[61824]: DEBUG nova.service [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] Join ServiceGroup membership for this service compute {{(pid=61824) start /opt/stack/nova/nova/service.py:203}} [ 476.617850] env[61824]: DEBUG nova.servicegroup.drivers.db [None req-b889cf64-4d73-45d3-adbd-3b3bb8a0017d None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61824) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 512.695597] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Acquiring lock "2646b1f8-9785-4517-a1d3-2588cd819f2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.695916] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Lock "2646b1f8-9785-4517-a1d3-2588cd819f2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.200223] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 513.451406] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.451723] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.740473] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.741445] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.742799] env[61824]: INFO nova.compute.claims [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.946814] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.946814] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.955071] env[61824]: DEBUG nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.271107] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Acquiring lock "c9690979-6f9e-4184-80db-b8f9659aa0c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.271329] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Lock "c9690979-6f9e-4184-80db-b8f9659aa0c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.450305] env[61824]: DEBUG nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.486095] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.777879] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.830103] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.830896] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.900161] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f778a7-2255-49c2-997c-bccf314e8fb9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.910104] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2db97fd-da44-4a0f-b02d-c6f4e06af167 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.965749] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667d1142-66c0-4bfc-8cac-8b8fc50af594 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.985414] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5526a445-17dc-4cae-9e98-d73995b146d8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.997093] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.008961] env[61824]: DEBUG nova.compute.provider_tree [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 515.082151] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Acquiring lock "be3227f2-3fe1-4767-90a3-55ee57c05df0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.082151] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Lock "be3227f2-3fe1-4767-90a3-55ee57c05df0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.308811] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.335550] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.512438] env[61824]: DEBUG nova.scheduler.client.report [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 515.578020] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "8a78dbd9-6ac2-4652-85e7-7ee59be346f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.578268] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "8a78dbd9-6ac2-4652-85e7-7ee59be346f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.587674] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.865049] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.021231] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.280s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.022041] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 516.024339] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.538s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.026104] env[61824]: INFO nova.compute.claims [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 516.080502] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.116682] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.533608] env[61824]: DEBUG nova.compute.utils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 516.538707] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 516.540580] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 516.616064] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.868081] env[61824]: DEBUG nova.policy [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef8264e4d6744fedb0f4d6181dd7b925', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b33b3cab9634d9aaed61eddb78f1e71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 517.047993] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 517.200907] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa957452-9ec8-42e7-815b-aff7cbe46dfc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.210954] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a5f336-d2f2-4cb1-88e5-ac0607eb3507 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.251695] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4433b42c-33c9-4670-84a4-3bddbd78d559 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.259810] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520dcb69-6e62-468e-b9a5-0702194a8527 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.285286] env[61824]: DEBUG nova.compute.provider_tree [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.790494] env[61824]: DEBUG nova.scheduler.client.report [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.927032] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "ab91f88f-6e87-405f-81ab-bf00be2003a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.927032] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "ab91f88f-6e87-405f-81ab-bf00be2003a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.964901] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Successfully created port: 6deea1d5-d0fa-4046-b15f-4d5a06f70d57 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 518.064027] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 518.104509] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 518.104752] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 518.104929] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 518.105242] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 518.105400] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 518.105543] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 518.105746] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 518.105932] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 518.106376] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 518.106546] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 518.106712] env[61824]: DEBUG nova.virt.hardware [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 518.107715] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4204785a-46fb-49cf-b3bb-994216ebaa88 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.116339] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7197a26d-c29f-467d-8e76-f419c3efedc6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.133563] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505f2ac6-e066-4dff-9153-3c31df8b40bd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.148675] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.148675] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.295273] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.296151] env[61824]: DEBUG nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 518.301260] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.305s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.302985] env[61824]: INFO nova.compute.claims [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.430365] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.652129] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.809705] env[61824]: DEBUG nova.compute.utils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.812978] env[61824]: DEBUG nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Not allocating networking since 'none' was specified. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 518.970791] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.181179] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.314704] env[61824]: DEBUG nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 519.466226] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ce8ac4-9533-4e6b-81ce-22350b7c5612 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.474425] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45dbf0b1-d7bb-4500-a457-9b04a4b2b182 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.515664] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec9129c-18f7-4ebd-ac34-b6ee3f875dea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.524412] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1d3391-e249-462d-8fce-1653e340eba6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.539313] env[61824]: DEBUG nova.compute.provider_tree [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.042378] env[61824]: DEBUG nova.scheduler.client.report [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.310251] env[61824]: ERROR nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. [ 520.310251] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 520.310251] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.310251] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 520.310251] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 520.310251] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 520.310251] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 520.310251] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 520.310251] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.310251] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 520.310251] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.310251] env[61824]: ERROR nova.compute.manager raise self.value [ 520.310251] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 520.310251] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 520.310251] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.310251] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 520.310697] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.310697] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 520.310697] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. [ 520.310697] env[61824]: ERROR nova.compute.manager [ 520.310697] env[61824]: Traceback (most recent call last): [ 520.310697] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 520.310697] env[61824]: listener.cb(fileno) [ 520.310697] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.310697] env[61824]: result = function(*args, **kwargs) [ 520.310697] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.310697] env[61824]: return func(*args, **kwargs) [ 520.310697] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.310697] env[61824]: raise e [ 520.310697] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.310697] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 520.310697] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 520.310697] env[61824]: created_port_ids = self._update_ports_for_instance( [ 520.310697] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 520.310697] env[61824]: with excutils.save_and_reraise_exception(): [ 520.310697] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.310697] env[61824]: self.force_reraise() [ 520.310697] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.310697] env[61824]: raise self.value [ 520.310697] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 520.310697] env[61824]: updated_port = self._update_port( [ 520.310697] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.310697] env[61824]: _ensure_no_port_binding_failure(port) [ 520.310697] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.310697] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 520.311516] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. [ 520.311516] env[61824]: Removing descriptor: 15 [ 520.315456] env[61824]: ERROR nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Traceback (most recent call last): [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] yield resources [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self.driver.spawn(context, instance, image_meta, [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] vm_ref = self.build_virtual_machine(instance, [ 520.315456] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] vif_infos = vmwarevif.get_vif_info(self._session, [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] for vif in network_info: [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return self._sync_wrapper(fn, *args, **kwargs) [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self.wait() [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self[:] = self._gt.wait() [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return self._exit_event.wait() [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 520.315893] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] result = hub.switch() [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return self.greenlet.switch() [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] result = function(*args, **kwargs) [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return func(*args, **kwargs) [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] raise e [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] nwinfo = self.network_api.allocate_for_instance( [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] created_port_ids = self._update_ports_for_instance( [ 520.316233] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] with excutils.save_and_reraise_exception(): [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self.force_reraise() [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] raise self.value [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] updated_port = self._update_port( [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] _ensure_no_port_binding_failure(port) [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] raise exception.PortBindingFailed(port_id=port['id']) [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] nova.exception.PortBindingFailed: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. [ 520.316563] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] [ 520.316901] env[61824]: INFO nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Terminating instance [ 520.322153] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Acquiring lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 520.322153] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Acquired lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.322278] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 520.328859] env[61824]: DEBUG nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 520.368956] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.368956] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.369144] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.369554] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.369554] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.369709] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.370194] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.370590] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.370809] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.371017] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.371232] env[61824]: DEBUG nova.virt.hardware [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.372144] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46f7452-173b-4dbd-a923-8a6e048a4b29 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.382065] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b87e00-5b31-4881-b88a-3d2fbc7845e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.397254] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 520.407230] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 520.407559] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05e3c7d6-b9d6-429b-8f79-c9e158329633 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.421665] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Created folder: OpenStack in parent group-v4. [ 520.421665] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Creating folder: Project (4d94673ee4d245c3b59ee3548fc99e24). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 520.421665] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddd93063-02ef-466a-81b4-136093342034 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.444222] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Created folder: Project (4d94673ee4d245c3b59ee3548fc99e24) in parent group-v274074. [ 520.444222] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Creating folder: Instances. Parent ref: group-v274075. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 520.444222] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce626920-104e-4a50-8830-4724c679c063 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.453770] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Created folder: Instances in parent group-v274075. [ 520.453770] env[61824]: DEBUG oslo.service.loopingcall [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.453770] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 520.453770] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38eb205c-a5b6-4903-add1-746ee400e8b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.478106] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 520.478106] env[61824]: value = "task-1275287" [ 520.478106] env[61824]: _type = "Task" [ 520.478106] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.487779] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275287, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.554523] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.555079] env[61824]: DEBUG nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.564266] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.253s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.570076] env[61824]: INFO nova.compute.claims [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.883130] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 520.993230] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275287, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.077632] env[61824]: DEBUG nova.compute.utils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.084869] env[61824]: DEBUG nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Not allocating networking since 'none' was specified. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 521.145417] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.392983] env[61824]: DEBUG nova.compute.manager [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Received event network-changed-6deea1d5-d0fa-4046-b15f-4d5a06f70d57 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 521.393182] env[61824]: DEBUG nova.compute.manager [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Refreshing instance network info cache due to event network-changed-6deea1d5-d0fa-4046-b15f-4d5a06f70d57. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 521.393401] env[61824]: DEBUG oslo_concurrency.lockutils [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] Acquiring lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.493599] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275287, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.580223] env[61824]: DEBUG nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.648927] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Releasing lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.650982] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 521.651167] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 521.652819] env[61824]: DEBUG oslo_concurrency.lockutils [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] Acquired lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.654017] env[61824]: DEBUG nova.network.neutron [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Refreshing network info cache for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 521.655366] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e145fa6-445e-4a65-9e1e-5b566050cad3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.670699] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b31a9b-9d1f-4850-9cbe-3ffc1f6d1175 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.701024] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2646b1f8-9785-4517-a1d3-2588cd819f2b could not be found. [ 521.701024] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 521.701024] env[61824]: INFO nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 521.701313] env[61824]: DEBUG oslo.service.loopingcall [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 521.704796] env[61824]: DEBUG nova.compute.manager [-] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 521.704845] env[61824]: DEBUG nova.network.neutron [-] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 521.755197] env[61824]: DEBUG nova.network.neutron [-] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 521.796400] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894dce92-e7c5-4123-a8c2-e558a0a8ebd2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.806630] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcb2b73-7852-4649-80af-a41a385bd65d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.840538] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca995c4-fe3c-4d41-9670-adf5e16ef888 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.848337] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6192699-def5-4ca7-af26-84ef0abfc53d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.863740] env[61824]: DEBUG nova.compute.provider_tree [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.998705] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275287, 'name': CreateVM_Task, 'duration_secs': 1.287903} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.998953] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 522.002428] env[61824]: DEBUG oslo_vmware.service [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8127a1-5ff0-47be-9e96-d00e32e6e8e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.009224] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.009386] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.011573] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 522.013255] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1cf7291-e3e3-4a1f-967d-797975deaba5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.017245] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 522.017245] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52bff715-2828-0b0a-c7dd-c095a40a6a71" [ 522.017245] env[61824]: _type = "Task" [ 522.017245] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.024791] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bff715-2828-0b0a-c7dd-c095a40a6a71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.218618] env[61824]: DEBUG nova.network.neutron [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 522.261251] env[61824]: DEBUG nova.network.neutron [-] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.366348] env[61824]: DEBUG nova.scheduler.client.report [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.535081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.535081] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 522.535081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.535081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.535379] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 522.535379] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0dff1805-8457-4d0c-b06a-259d50f70e70 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.553451] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 522.554113] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 522.556399] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5639ed0a-aa0f-40af-ae1f-2e07844b47b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.569021] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72a4c1e9-dfc8-49fd-a4c1-8777bd14b32d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.575346] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 522.575346] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5227a6de-0e72-6d26-66a6-afc02c21b67a" [ 522.575346] env[61824]: _type = "Task" [ 522.575346] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.583581] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5227a6de-0e72-6d26-66a6-afc02c21b67a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.594609] env[61824]: DEBUG nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.598374] env[61824]: DEBUG nova.network.neutron [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.648842] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.648842] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.649446] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.649746] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.651352] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.651352] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.651352] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.651352] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.651352] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.651743] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.651743] env[61824]: DEBUG nova.virt.hardware [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.651806] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddd037a-7826-441e-9274-41ad0eb7b6b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.665801] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e82a858-e942-4723-88f6-7cf0a7f1b0b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.693264] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 522.699860] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Creating folder: Project (6fcf9d61c31d41dda107e8c82df39a9b). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.700195] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-31d7455b-9dba-47f7-bde3-4eb208380278 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.712104] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Created folder: Project (6fcf9d61c31d41dda107e8c82df39a9b) in parent group-v274074. [ 522.712752] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Creating folder: Instances. Parent ref: group-v274078. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 522.713685] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c733ba7-ab0a-4bc0-8cb8-7e5cde16bf29 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.722536] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Created folder: Instances in parent group-v274078. [ 522.722779] env[61824]: DEBUG oslo.service.loopingcall [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.723274] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 522.723274] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5ce30e99-e563-4983-95ad-236ebe6f4080 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.743318] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 522.743318] env[61824]: value = "task-1275290" [ 522.743318] env[61824]: _type = "Task" [ 522.743318] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.751678] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275290, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.764869] env[61824]: INFO nova.compute.manager [-] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Took 1.06 seconds to deallocate network for instance. [ 522.767876] env[61824]: DEBUG nova.compute.claims [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 522.768107] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.875109] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.875109] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.876230] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.011s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.884538] env[61824]: INFO nova.compute.claims [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.086033] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Preparing fetch location {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 523.086033] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Creating directory with path [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 523.086033] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a9257aef-1398-433b-a280-953ec76adf37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.104892] env[61824]: DEBUG oslo_concurrency.lockutils [req-0b554bff-71a1-4204-a7c5-1de28a80e1b0 req-876b58d9-b32b-49a7-9657-af44934b45d7 service nova] Releasing lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.107680] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Created directory with path [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 523.108095] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Fetch image to [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 523.108405] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Downloading image file data 11b2670a-52f3-4a0e-8363-54ca7a7f5264 to [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk on the data store datastore2 {{(pid=61824) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 523.110045] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf630e0-d8c8-4f31-9f53-143e9858de48 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.121164] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa671d11-48f2-4bc1-b3f8-c81413408639 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.133530] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa9f381-154d-4011-b510-01057b6695f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.172858] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db58825b-0722-4a42-96ba-770bf413fd38 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.180440] env[61824]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-e4536a46-5014-44e9-b121-35fc2e98e707 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.263116] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275290, 'name': CreateVM_Task, 'duration_secs': 0.286363} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.263116] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 523.263116] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.263116] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.263116] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 523.263116] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39d8cc77-33ab-4d65-8d61-0c41ca327319 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.269172] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 523.269172] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5266c03a-fc3f-f130-ef56-d0c57d61db48" [ 523.269172] env[61824]: _type = "Task" [ 523.269172] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.271373] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Downloading image file data 11b2670a-52f3-4a0e-8363-54ca7a7f5264 to the data store datastore2 {{(pid=61824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 523.288407] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.288654] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 523.288907] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.347735] env[61824]: DEBUG oslo_vmware.rw_handles [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 523.413059] env[61824]: DEBUG nova.compute.utils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.418936] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.419222] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 523.812979] env[61824]: DEBUG nova.policy [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b729b46bd6a4c7ab2ddbd78a23c474c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '08a921d66ba440e6a0f5b1b0578d99d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.916726] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 524.005031] env[61824]: DEBUG oslo_vmware.rw_handles [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Completed reading data from the image iterator. {{(pid=61824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 524.005031] env[61824]: DEBUG oslo_vmware.rw_handles [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Closing write handle for https://esx7c2n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 524.141717] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Downloaded image file data 11b2670a-52f3-4a0e-8363-54ca7a7f5264 to vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk on the data store datastore2 {{(pid=61824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 524.142977] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Caching image {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 524.143233] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Copying Virtual Disk [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk to [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 524.143573] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbfb3aaf-cff3-459d-8a8f-6377e3b2485f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.156708] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 524.156708] env[61824]: value = "task-1275291" [ 524.156708] env[61824]: _type = "Task" [ 524.156708] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.178976] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.234151] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c4cb03-1aff-4f37-92e0-3f65af606580 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.242470] env[61824]: DEBUG nova.compute.manager [req-165f9381-e7d3-48f3-92b0-9cfc5ea4fc88 req-574a3487-81b4-48f5-9db8-b2e4e36b76dc service nova] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Received event network-vif-deleted-6deea1d5-d0fa-4046-b15f-4d5a06f70d57 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 524.249781] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79732bbf-b087-4953-aaa1-ace0b7e57fda {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.291884] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4804387-707a-402b-81db-0cbfd533eaf5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.300020] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b178b5b-2867-47a2-99ab-19218f1deeb4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.315527] env[61824]: DEBUG nova.compute.provider_tree [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.672749] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275291, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.819155] env[61824]: DEBUG nova.scheduler.client.report [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.931033] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.968169] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.968408] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.968558] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.969388] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.969388] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.969388] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.969388] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.969388] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.969656] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.969705] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.969838] env[61824]: DEBUG nova.virt.hardware [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.970734] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d77de4-e714-42c1-8872-4f8ccab0080a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.980540] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc20d4af-4af9-4ac6-950c-5f0ef07f3a5b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.178108] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672701} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.179199] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Copied Virtual Disk [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk to [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 525.179199] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Deleting the datastore file [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 525.179199] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f648ec3-9eee-4dec-b12f-d418691824c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.189708] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 525.189708] env[61824]: value = "task-1275292" [ 525.189708] env[61824]: _type = "Task" [ 525.189708] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.202182] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275292, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.326384] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.326729] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.329806] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.213s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.331352] env[61824]: INFO nova.compute.claims [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.705487] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275292, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025583} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.707085] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 525.707085] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Moving file from [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7/11b2670a-52f3-4a0e-8363-54ca7a7f5264 to [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264. {{(pid=61824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 525.707085] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-cd220c3e-f852-4252-92f6-dec11c2e2166 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.714522] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 525.714522] env[61824]: value = "task-1275293" [ 525.714522] env[61824]: _type = "Task" [ 525.714522] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.725599] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275293, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.783578] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Acquiring lock "2fb76ee4-a42b-4de6-9889-b9a89cf16533" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.783831] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Lock "2fb76ee4-a42b-4de6-9889-b9a89cf16533" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.831424] env[61824]: DEBUG nova.compute.utils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.832853] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.833515] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 526.074229] env[61824]: DEBUG nova.policy [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fab4d7e1cb4da9a54c55cc9762c329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d19b5b4169b4a55b1d14216da62a659', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.230838] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275293, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023801} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.235281] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] File moved {{(pid=61824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 526.235641] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Cleaning up location [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 526.236403] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Deleting the datastore file [datastore2] vmware_temp/66ecfc0b-b1da-4b35-8075-22ea82303ae7 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 526.236403] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff3fd9d4-9e50-43fe-84f1-178c8bf55c00 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.248706] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 526.248706] env[61824]: value = "task-1275294" [ 526.248706] env[61824]: _type = "Task" [ 526.248706] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.261160] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275294, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.289781] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.324616] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Successfully created port: fb461612-dcac-4cb1-80b6-9413277bacb1 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.337790] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.652119] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a3dc96-82eb-456c-a709-b70d0149cd02 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.661288] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e99519-36e0-4f11-9ff9-708b7471d933 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.699519] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d291ba-f425-4a79-83e9-3d1ed5520649 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.707436] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9071f425-7a00-4391-bbc0-21c8a2b3270f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.721871] env[61824]: DEBUG nova.compute.provider_tree [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.757527] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275294, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025936} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.757812] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 526.758510] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4becf06-5c50-4550-9305-0af0dcaff254 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.764536] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 526.764536] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]526ea7b7-c571-91fc-d849-271dd356be20" [ 526.764536] env[61824]: _type = "Task" [ 526.764536] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 526.777795] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526ea7b7-c571-91fc-d849-271dd356be20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.827776] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.227335] env[61824]: DEBUG nova.scheduler.client.report [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.282016] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526ea7b7-c571-91fc-d849-271dd356be20, 'name': SearchDatastore_Task, 'duration_secs': 0.009055} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.282016] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.282016] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468/90a21ec5-7ccf-4d37-9ec9-19ba2bfff468.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 527.282016] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.282268] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 527.282268] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6be33e53-2cf6-46ad-a875-5fcb96ccc34d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.285362] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d935af01-8cde-4038-8205-04497e963679 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.297020] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 527.297020] env[61824]: value = "task-1275295" [ 527.297020] env[61824]: _type = "Task" [ 527.297020] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.299186] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 527.299577] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 527.304978] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12374ad9-5210-4c2b-8ef9-30418a9e6dc4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.315425] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275295, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.317331] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 527.317331] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52ee9cc4-08ae-e654-b4de-634779565fc1" [ 527.317331] env[61824]: _type = "Task" [ 527.317331] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.329259] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ee9cc4-08ae-e654-b4de-634779565fc1, 'name': SearchDatastore_Task, 'duration_secs': 0.007405} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.330083] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4966333d-b190-454b-b910-96d87d9af1a3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.337229] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 527.337229] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52cba26a-a7da-b8de-b751-3c6c3fe160a3" [ 527.337229] env[61824]: _type = "Task" [ 527.337229] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.346203] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52cba26a-a7da-b8de-b751-3c6c3fe160a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.356587] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.394482] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.395139] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.399053] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.399053] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.399053] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.399053] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.399053] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.399339] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.399339] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.399339] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.399339] env[61824]: DEBUG nova.virt.hardware [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.399339] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e21d024-34e3-4a75-a1e3-6787be9e3f0e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.409397] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973a9b78-de2b-4547-ac5f-299aa23211ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.619444] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.621063] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 527.621063] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 527.621063] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Rebuilding the list of instances to heal {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 527.738436] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.739071] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.742503] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.127s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.748146] env[61824]: INFO nova.compute.claims [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.807357] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275295, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468224} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.807468] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468/90a21ec5-7ccf-4d37-9ec9-19ba2bfff468.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 527.808896] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 527.808896] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42f4f0cb-d0e3-4f24-ad97-ee8367009ab3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.817892] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 527.817892] env[61824]: value = "task-1275296" [ 527.817892] env[61824]: _type = "Task" [ 527.817892] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.833534] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275296, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.847256] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52cba26a-a7da-b8de-b751-3c6c3fe160a3, 'name': SearchDatastore_Task, 'duration_secs': 0.008417} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 527.847431] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.847906] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 527.848060] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ffdcf7a-9d6c-403b-a49e-895d5a4a105a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.856933] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 527.856933] env[61824]: value = "task-1275297" [ 527.856933] env[61824]: _type = "Task" [ 527.856933] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 527.866958] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275297, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.133017] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.133017] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.133017] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.133017] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.133017] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.133017] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 528.133209] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Didn't find any instances for network info cache update. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 528.133209] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.133209] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.133209] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.133209] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.133209] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.133455] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_power_states {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.255220] env[61824]: DEBUG nova.compute.utils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.262609] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.262747] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 528.333122] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275296, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066415} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.333399] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 528.334309] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b49a0c2-c127-4f69-9464-7f644519a864 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.359871] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468/90a21ec5-7ccf-4d37-9ec9-19ba2bfff468.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 528.360381] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c098b3c-497d-4f03-9afb-a6002b3e8dc6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.391224] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275297, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463866} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.392653] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 528.392873] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 528.393376] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 528.393376] env[61824]: value = "task-1275298" [ 528.393376] env[61824]: _type = "Task" [ 528.393376] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.393784] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b99aeb7-8f62-4ac9-a1aa-490b93c297fe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.407645] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275298, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.410447] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 528.410447] env[61824]: value = "task-1275299" [ 528.410447] env[61824]: _type = "Task" [ 528.410447] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.449161] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Successfully created port: c88594d9-61fa-4595-bb48-f998560bf820 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.502105] env[61824]: DEBUG nova.policy [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41c38d18848d4b7fa66a77cee0fe8ef9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e17ae96cb12f416ba89f0a635550751b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.644062] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Getting list of instances from cluster (obj){ [ 528.644062] env[61824]: value = "domain-c8" [ 528.644062] env[61824]: _type = "ClusterComputeResource" [ 528.644062] env[61824]: } {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 528.646766] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4c8f27-7106-47a8-8fa8-8578c3e583a0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.663191] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Got total of 2 instances {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 528.663191] env[61824]: WARNING nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] While synchronizing instance power states, found 7 instances in the database and 2 instances on the hypervisor. [ 528.663191] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid 2646b1f8-9785-4517-a1d3-2588cd819f2b {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 528.665128] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468 {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 528.665128] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid ff1dac51-f008-4218-bd4c-f0104657a2ee {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 528.665128] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid c9690979-6f9e-4184-80db-b8f9659aa0c2 {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 528.670470] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946 {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 528.670470] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid be3227f2-3fe1-4767-90a3-55ee57c05df0 {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 528.670470] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid 8a78dbd9-6ac2-4652-85e7-7ee59be346f4 {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 528.670470] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "2646b1f8-9785-4517-a1d3-2588cd819f2b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.670470] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.670470] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.671920] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "c9690979-6f9e-4184-80db-b8f9659aa0c2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.671920] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.671920] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "be3227f2-3fe1-4767-90a3-55ee57c05df0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.671920] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "8a78dbd9-6ac2-4652-85e7-7ee59be346f4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.671920] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.678288] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 528.678288] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 528.763905] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.906687] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275298, 'name': ReconfigVM_Task, 'duration_secs': 0.293584} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.909906] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468/90a21ec5-7ccf-4d37-9ec9-19ba2bfff468.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 528.911503] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0f109a0-6138-4b9b-8a5f-a032d4193290 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.926639] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 528.926639] env[61824]: value = "task-1275300" [ 528.926639] env[61824]: _type = "Task" [ 528.926639] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.926639] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275299, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095093} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.929509] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 528.933546] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-878209b9-c812-4966-8e9b-4b50ea4fb0a3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.961814] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 528.969614] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dde24b1b-6aec-41c9-8482-1b69d1df5ee2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.990740] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 528.990740] env[61824]: value = "task-1275301" [ 528.990740] env[61824]: _type = "Task" [ 528.990740] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.002162] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275301, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.010245] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Acquiring lock "abc541cd-4dfc-4071-9b3d-2ec220f225ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.010245] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Lock "abc541cd-4dfc-4071-9b3d-2ec220f225ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.064584] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18a6c84-aa9c-4b0d-b575-c1240bcd9da2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.073721] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92d97b3-b58f-46d8-a8d0-2874216af371 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.108578] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb00680-6756-4f67-81bc-99b0d0da9c78 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.117395] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea028db-550c-4d88-b297-da16cc17ef90 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.133439] env[61824]: DEBUG nova.compute.provider_tree [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.184487] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.439281] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275300, 'name': Rename_Task, 'duration_secs': 0.157584} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.439708] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 529.440008] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-866f8df2-95d8-4dc5-95b7-9f9ed93e510c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.446563] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 529.446563] env[61824]: value = "task-1275302" [ 529.446563] env[61824]: _type = "Task" [ 529.446563] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.456021] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275302, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.502889] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275301, 'name': ReconfigVM_Task, 'duration_secs': 0.290364} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.502889] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Reconfigured VM instance instance-00000003 to attach disk [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 529.507019] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ed07c8c-f3fb-4471-b099-576247443eb2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.514567] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 529.514567] env[61824]: value = "task-1275303" [ 529.514567] env[61824]: _type = "Task" [ 529.514567] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.527526] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275303, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.636560] env[61824]: DEBUG nova.scheduler.client.report [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.787349] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.820024] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.820024] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.820024] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.820227] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.820437] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.820755] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.821124] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.821487] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.821705] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.821903] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.822203] env[61824]: DEBUG nova.virt.hardware [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.823828] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437d734b-7f53-4011-9c8e-cb15d8ea82b4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.833469] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b7ff49-2e1c-4800-ba3a-57dbf39a3e36 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.962726] env[61824]: DEBUG oslo_vmware.api [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275302, 'name': PowerOnVM_Task, 'duration_secs': 0.440572} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.963715] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Successfully created port: 25203994-eda5-4335-95fc-7fa68d2320d9 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.965736] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 529.965828] env[61824]: INFO nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Took 9.64 seconds to spawn the instance on the hypervisor. [ 529.971995] env[61824]: DEBUG nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 529.972840] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f7f657-498d-48df-9145-94b77a32a785 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.028865] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275303, 'name': Rename_Task, 'duration_secs': 0.143979} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.028865] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 530.028865] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ea292300-d53c-47e5-a76d-a365b01cf3b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.033999] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 530.033999] env[61824]: value = "task-1275304" [ 530.033999] env[61824]: _type = "Task" [ 530.033999] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.048279] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275304, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.144582] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.145165] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.147693] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.177s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.149303] env[61824]: INFO nova.compute.claims [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.504544] env[61824]: INFO nova.compute.manager [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Took 16.04 seconds to build instance. [ 530.546970] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275304, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.656464] env[61824]: DEBUG nova.compute.utils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.659446] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.659446] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.778308] env[61824]: DEBUG nova.policy [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6a2f1ffe8934519acdcaca77e2a55d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e51dd2f578344aa8afa5d6beccb54f1e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.010121] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4f93228-2726-441d-913a-9b6878dd3d63 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.558s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.011786] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.343s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.011786] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] During sync_power_state the instance has a pending task (spawning). Skip. [ 531.011786] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.054842] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275304, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.161119] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.413013] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c824481-3c89-455a-a682-1075dbe93a77 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.423538] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f48c752-c8aa-42d3-88a3-17954334fa6d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.456674] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1417be5c-ee51-4bed-9d00-d4d2867b9150 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.464677] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a68074c-162e-45dc-afee-30425bb20c4e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.485569] env[61824]: DEBUG nova.compute.provider_tree [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.513542] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 531.524807] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Acquiring lock "e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.525167] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Lock "e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.551160] env[61824]: DEBUG oslo_vmware.api [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275304, 'name': PowerOnVM_Task, 'duration_secs': 1.160139} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.551160] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 531.551160] env[61824]: INFO nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Took 8.95 seconds to spawn the instance on the hypervisor. [ 531.551160] env[61824]: DEBUG nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 531.551160] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8c8b76-ecb0-4faf-ae1a-85d454a52135 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.994623] env[61824]: DEBUG nova.scheduler.client.report [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.048383] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.079074] env[61824]: INFO nova.compute.manager [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Took 17.12 seconds to build instance. [ 532.098153] env[61824]: ERROR nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. [ 532.098153] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.098153] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.098153] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.098153] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.098153] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.098153] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.098153] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.098153] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.098153] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 532.098153] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.098153] env[61824]: ERROR nova.compute.manager raise self.value [ 532.098153] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.098153] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.098153] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.098153] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.098666] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.098666] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.098666] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. [ 532.098666] env[61824]: ERROR nova.compute.manager [ 532.098666] env[61824]: Traceback (most recent call last): [ 532.098666] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.098666] env[61824]: listener.cb(fileno) [ 532.098666] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.098666] env[61824]: result = function(*args, **kwargs) [ 532.098666] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.098666] env[61824]: return func(*args, **kwargs) [ 532.098666] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.098666] env[61824]: raise e [ 532.098666] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.098666] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 532.098666] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.098666] env[61824]: created_port_ids = self._update_ports_for_instance( [ 532.098666] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.098666] env[61824]: with excutils.save_and_reraise_exception(): [ 532.098666] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.098666] env[61824]: self.force_reraise() [ 532.098666] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.098666] env[61824]: raise self.value [ 532.098666] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.098666] env[61824]: updated_port = self._update_port( [ 532.098666] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.098666] env[61824]: _ensure_no_port_binding_failure(port) [ 532.098666] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.098666] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.099543] env[61824]: nova.exception.PortBindingFailed: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. [ 532.099543] env[61824]: Removing descriptor: 15 [ 532.099543] env[61824]: ERROR nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Traceback (most recent call last): [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] yield resources [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self.driver.spawn(context, instance, image_meta, [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.099543] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] vm_ref = self.build_virtual_machine(instance, [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] for vif in network_info: [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return self._sync_wrapper(fn, *args, **kwargs) [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self.wait() [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self[:] = self._gt.wait() [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return self._exit_event.wait() [ 532.099877] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] result = hub.switch() [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return self.greenlet.switch() [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] result = function(*args, **kwargs) [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return func(*args, **kwargs) [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] raise e [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] nwinfo = self.network_api.allocate_for_instance( [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.100270] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] created_port_ids = self._update_ports_for_instance( [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] with excutils.save_and_reraise_exception(): [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self.force_reraise() [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] raise self.value [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] updated_port = self._update_port( [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] _ensure_no_port_binding_failure(port) [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.100656] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] raise exception.PortBindingFailed(port_id=port['id']) [ 532.101054] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] nova.exception.PortBindingFailed: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. [ 532.101054] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] [ 532.101054] env[61824]: INFO nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Terminating instance [ 532.107467] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Acquiring lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.107467] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Acquired lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.107467] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 532.177560] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.208155] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.208520] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.208746] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.209775] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.209775] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.209775] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.209775] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.209775] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.209991] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.210272] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.211471] env[61824]: DEBUG nova.virt.hardware [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.211471] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f30618a1-5f11-4924-89c8-ed84c6ed5706 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.221841] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224c2088-bafd-4580-8ac5-f04d971d8292 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.251729] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Successfully created port: f7dfe812-b4b9-4bb3-8b72-d398ba14bacc {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.499733] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.352s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.500359] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.503246] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.322s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.508466] env[61824]: INFO nova.compute.claims [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.583554] env[61824]: DEBUG oslo_concurrency.lockutils [None req-47984310-51f2-4d42-9c06-c829a26a9fbe tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.638s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.584919] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.916s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.585268] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] During sync_power_state the instance has a pending task (spawning). Skip. [ 532.585337] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.690355] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.701194] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Acquiring lock "a062288b-e785-4dec-98f2-358b54f3d327" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.701887] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Lock "a062288b-e785-4dec-98f2-358b54f3d327" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.881796] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.013156] env[61824]: DEBUG nova.compute.utils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.018091] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 533.018091] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 533.088246] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.205190] env[61824]: DEBUG nova.policy [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632dafc515e34f318f6f4b43bd1a62aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '783bb630c83943f286b3d019c2832ba1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.386355] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Releasing lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.387207] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 533.387207] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 533.387347] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a351e88-17b6-4522-80fc-82e26db11b55 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.400490] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc69c59-3286-43d5-ad55-e4b908660293 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.429020] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9690979-6f9e-4184-80db-b8f9659aa0c2 could not be found. [ 533.429315] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 533.429528] env[61824]: INFO nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 533.430434] env[61824]: DEBUG oslo.service.loopingcall [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.432655] env[61824]: DEBUG nova.compute.manager [-] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.432728] env[61824]: DEBUG nova.network.neutron [-] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 533.475106] env[61824]: ERROR nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. [ 533.475106] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.475106] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.475106] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.475106] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.475106] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.475106] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.475106] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.475106] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.475106] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 533.475106] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.475106] env[61824]: ERROR nova.compute.manager raise self.value [ 533.475106] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.475106] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.475106] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.475106] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.476126] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.476126] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.476126] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. [ 533.476126] env[61824]: ERROR nova.compute.manager [ 533.476126] env[61824]: Traceback (most recent call last): [ 533.476126] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.476126] env[61824]: listener.cb(fileno) [ 533.476126] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.476126] env[61824]: result = function(*args, **kwargs) [ 533.476126] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.476126] env[61824]: return func(*args, **kwargs) [ 533.476126] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.476126] env[61824]: raise e [ 533.476126] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.476126] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 533.476126] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.476126] env[61824]: created_port_ids = self._update_ports_for_instance( [ 533.476126] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.476126] env[61824]: with excutils.save_and_reraise_exception(): [ 533.476126] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.476126] env[61824]: self.force_reraise() [ 533.476126] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.476126] env[61824]: raise self.value [ 533.476126] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.476126] env[61824]: updated_port = self._update_port( [ 533.476126] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.476126] env[61824]: _ensure_no_port_binding_failure(port) [ 533.476126] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.476126] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.478823] env[61824]: nova.exception.PortBindingFailed: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. [ 533.478823] env[61824]: Removing descriptor: 17 [ 533.478823] env[61824]: ERROR nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Traceback (most recent call last): [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] yield resources [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self.driver.spawn(context, instance, image_meta, [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.478823] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] vm_ref = self.build_virtual_machine(instance, [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] for vif in network_info: [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return self._sync_wrapper(fn, *args, **kwargs) [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self.wait() [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self[:] = self._gt.wait() [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return self._exit_event.wait() [ 533.479284] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] result = hub.switch() [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return self.greenlet.switch() [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] result = function(*args, **kwargs) [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return func(*args, **kwargs) [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] raise e [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] nwinfo = self.network_api.allocate_for_instance( [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.479638] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] created_port_ids = self._update_ports_for_instance( [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] with excutils.save_and_reraise_exception(): [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self.force_reraise() [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] raise self.value [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] updated_port = self._update_port( [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] _ensure_no_port_binding_failure(port) [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.479984] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] raise exception.PortBindingFailed(port_id=port['id']) [ 533.480344] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] nova.exception.PortBindingFailed: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. [ 533.480344] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] [ 533.480344] env[61824]: INFO nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Terminating instance [ 533.483101] env[61824]: DEBUG nova.network.neutron [-] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.487968] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.488372] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquired lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.488576] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.525558] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.613263] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.638099] env[61824]: DEBUG nova.compute.manager [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Received event network-changed-fb461612-dcac-4cb1-80b6-9413277bacb1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.639021] env[61824]: DEBUG nova.compute.manager [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Refreshing instance network info cache due to event network-changed-fb461612-dcac-4cb1-80b6-9413277bacb1. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.639021] env[61824]: DEBUG oslo_concurrency.lockutils [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] Acquiring lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.639021] env[61824]: DEBUG oslo_concurrency.lockutils [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] Acquired lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.639249] env[61824]: DEBUG nova.network.neutron [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Refreshing network info cache for port fb461612-dcac-4cb1-80b6-9413277bacb1 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 533.817822] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0d4cc0-88d6-4357-acb6-836afa0ad43b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.826259] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5bd739-659f-450b-948d-f1802a2302a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.858506] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db76cc08-aa09-44a6-8e3c-c32458b8f86b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.868030] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5bae49e-f788-48de-9072-8ff48805ea88 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.885487] env[61824]: DEBUG nova.compute.provider_tree [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.988517] env[61824]: DEBUG nova.network.neutron [-] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.030590] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.237338] env[61824]: DEBUG nova.network.neutron [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.389243] env[61824]: DEBUG nova.scheduler.client.report [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.425996] env[61824]: INFO nova.compute.manager [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Rebuilding instance [ 534.471590] env[61824]: DEBUG nova.compute.manager [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 534.472700] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af26ded2-9ad2-4e4c-8cd6-75cd650d4878 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.486931] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.493947] env[61824]: INFO nova.compute.manager [-] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Took 1.06 seconds to deallocate network for instance. [ 534.494947] env[61824]: DEBUG nova.compute.claims [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 534.495171] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.546135] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.571949] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.572238] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.572371] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.572549] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.572693] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.572856] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.573088] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.573272] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.573437] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.574108] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.574108] env[61824]: DEBUG nova.virt.hardware [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.574661] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16375bb1-0a69-4c29-9c40-68d11221219d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.583927] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2b5fa0c-859f-4b61-a026-a6686d6a10df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.632795] env[61824]: DEBUG nova.network.neutron [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.709649] env[61824]: ERROR nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. [ 534.709649] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.709649] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.709649] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.709649] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.709649] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.709649] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.709649] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.709649] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.709649] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 534.709649] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.709649] env[61824]: ERROR nova.compute.manager raise self.value [ 534.709649] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.709649] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.709649] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.709649] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.710178] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.710178] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.710178] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. [ 534.710178] env[61824]: ERROR nova.compute.manager [ 534.710178] env[61824]: Traceback (most recent call last): [ 534.710178] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.710178] env[61824]: listener.cb(fileno) [ 534.710178] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.710178] env[61824]: result = function(*args, **kwargs) [ 534.710178] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.710178] env[61824]: return func(*args, **kwargs) [ 534.710178] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.710178] env[61824]: raise e [ 534.710178] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.710178] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 534.710178] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.710178] env[61824]: created_port_ids = self._update_ports_for_instance( [ 534.710178] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.710178] env[61824]: with excutils.save_and_reraise_exception(): [ 534.710178] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.710178] env[61824]: self.force_reraise() [ 534.710178] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.710178] env[61824]: raise self.value [ 534.710178] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.710178] env[61824]: updated_port = self._update_port( [ 534.710178] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.710178] env[61824]: _ensure_no_port_binding_failure(port) [ 534.710178] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.710178] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.710992] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. [ 534.710992] env[61824]: Removing descriptor: 18 [ 534.710992] env[61824]: ERROR nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Traceback (most recent call last): [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] yield resources [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self.driver.spawn(context, instance, image_meta, [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.710992] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] vm_ref = self.build_virtual_machine(instance, [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] for vif in network_info: [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return self._sync_wrapper(fn, *args, **kwargs) [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self.wait() [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self[:] = self._gt.wait() [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return self._exit_event.wait() [ 534.711340] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] result = hub.switch() [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return self.greenlet.switch() [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] result = function(*args, **kwargs) [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return func(*args, **kwargs) [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] raise e [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] nwinfo = self.network_api.allocate_for_instance( [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.711754] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] created_port_ids = self._update_ports_for_instance( [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] with excutils.save_and_reraise_exception(): [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self.force_reraise() [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] raise self.value [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] updated_port = self._update_port( [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] _ensure_no_port_binding_failure(port) [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.712172] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] raise exception.PortBindingFailed(port_id=port['id']) [ 534.712519] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] nova.exception.PortBindingFailed: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. [ 534.712519] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] [ 534.712519] env[61824]: INFO nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Terminating instance [ 534.714601] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Acquiring lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.714721] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Acquired lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.714866] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 534.762069] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Successfully created port: aea0ad08-56c4-45b0-87b6-c2b41e67f3b0 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.895347] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.895347] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.899172] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.131s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.984482] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 534.984894] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1b6888a-ca65-4d3f-8348-e98b2d7259c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.989413] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Releasing lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.989784] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.990011] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.991035] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23ae4af7-3b09-4063-9351-a8c9fe547c6e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.996063] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 534.996063] env[61824]: value = "task-1275305" [ 534.996063] env[61824]: _type = "Task" [ 534.996063] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.006326] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c390be9-bfc7-4c96-8f4b-badd9451cdfa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.023416] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.037014] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946 could not be found. [ 535.037330] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.037729] env[61824]: INFO nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Took 0.05 seconds to destroy the instance on the hypervisor. [ 535.037829] env[61824]: DEBUG oslo.service.loopingcall [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.037994] env[61824]: DEBUG nova.compute.manager [-] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.038255] env[61824]: DEBUG nova.network.neutron [-] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.089569] env[61824]: DEBUG nova.network.neutron [-] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.136070] env[61824]: DEBUG oslo_concurrency.lockutils [req-aa71a975-f962-413c-a262-a4fe31bc982e req-d5c13251-c0ea-41c6-918a-ef01e30df3fa service nova] Releasing lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.273804] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.404727] env[61824]: DEBUG nova.compute.utils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.406035] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.406211] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 535.515376] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275305, 'name': PowerOffVM_Task, 'duration_secs': 0.125638} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.517923] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 535.518232] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.519207] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bed6a4-613c-40c2-be4b-687572fa09d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.532101] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 535.532366] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c30d6d57-005c-44e3-bc02-6520604ce57c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.562073] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 535.562073] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 535.562073] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Deleting the datastore file [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 535.562073] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c078a57-f578-4a96-86fa-5765e12674e3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.569124] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 535.569124] env[61824]: value = "task-1275307" [ 535.569124] env[61824]: _type = "Task" [ 535.569124] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.583544] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275307, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.593667] env[61824]: DEBUG nova.network.neutron [-] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.652928] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad73db55-2558-4d6b-b19b-e3cd6fb51afa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.661586] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a109273-f30d-42e0-a40e-cc4bca4de86e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.695927] env[61824]: DEBUG nova.policy [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a81282e41a294fb9bf8524dd15704cfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92f9e969563247e1a0121e620ea5d2ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.698136] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44b8996-83d3-4595-a717-8f3b3dcb6a9a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.707597] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3449f097-d135-46a9-9e1d-14bfa2b7a842 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.720582] env[61824]: DEBUG nova.compute.provider_tree [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.725716] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.915565] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 536.083850] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275307, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094553} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.084126] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 536.084305] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 536.084911] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 536.099689] env[61824]: INFO nova.compute.manager [-] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Took 1.06 seconds to deallocate network for instance. [ 536.103867] env[61824]: DEBUG nova.compute.claims [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.105075] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.223499] env[61824]: DEBUG nova.scheduler.client.report [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.230930] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Releasing lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.231683] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 536.231926] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 536.235340] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1ea8595-3b84-4a15-87ac-48795afd6829 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.242450] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74c1acd-6642-4b90-a4d1-0533dfb8f87c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.269492] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance be3227f2-3fe1-4767-90a3-55ee57c05df0 could not be found. [ 536.269746] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 536.269932] env[61824]: INFO nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 536.270202] env[61824]: DEBUG oslo.service.loopingcall [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.270426] env[61824]: DEBUG nova.compute.manager [-] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.270557] env[61824]: DEBUG nova.network.neutron [-] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 536.386625] env[61824]: DEBUG nova.network.neutron [-] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.627793] env[61824]: ERROR nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. [ 536.627793] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.627793] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.627793] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.627793] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.627793] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.627793] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.627793] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.627793] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.627793] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 536.627793] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.627793] env[61824]: ERROR nova.compute.manager raise self.value [ 536.627793] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.627793] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.627793] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.627793] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.628547] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.628547] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.628547] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. [ 536.628547] env[61824]: ERROR nova.compute.manager [ 536.628547] env[61824]: Traceback (most recent call last): [ 536.628547] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.628547] env[61824]: listener.cb(fileno) [ 536.628547] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.628547] env[61824]: result = function(*args, **kwargs) [ 536.628547] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.628547] env[61824]: return func(*args, **kwargs) [ 536.628547] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.628547] env[61824]: raise e [ 536.628547] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.628547] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 536.628547] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.628547] env[61824]: created_port_ids = self._update_ports_for_instance( [ 536.628547] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.628547] env[61824]: with excutils.save_and_reraise_exception(): [ 536.628547] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.628547] env[61824]: self.force_reraise() [ 536.628547] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.628547] env[61824]: raise self.value [ 536.628547] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.628547] env[61824]: updated_port = self._update_port( [ 536.628547] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.628547] env[61824]: _ensure_no_port_binding_failure(port) [ 536.628547] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.628547] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.629430] env[61824]: nova.exception.PortBindingFailed: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. [ 536.629430] env[61824]: Removing descriptor: 19 [ 536.629430] env[61824]: ERROR nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Traceback (most recent call last): [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] yield resources [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self.driver.spawn(context, instance, image_meta, [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.629430] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] vm_ref = self.build_virtual_machine(instance, [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] for vif in network_info: [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return self._sync_wrapper(fn, *args, **kwargs) [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self.wait() [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self[:] = self._gt.wait() [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return self._exit_event.wait() [ 536.629985] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] result = hub.switch() [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return self.greenlet.switch() [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] result = function(*args, **kwargs) [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return func(*args, **kwargs) [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] raise e [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] nwinfo = self.network_api.allocate_for_instance( [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.630407] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] created_port_ids = self._update_ports_for_instance( [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] with excutils.save_and_reraise_exception(): [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self.force_reraise() [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] raise self.value [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] updated_port = self._update_port( [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] _ensure_no_port_binding_failure(port) [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.630941] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] raise exception.PortBindingFailed(port_id=port['id']) [ 536.631992] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] nova.exception.PortBindingFailed: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. [ 536.631992] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] [ 536.631992] env[61824]: INFO nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Terminating instance [ 536.634274] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.634274] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.634458] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.735154] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.836s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.735829] env[61824]: ERROR nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Traceback (most recent call last): [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self.driver.spawn(context, instance, image_meta, [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] vm_ref = self.build_virtual_machine(instance, [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.735829] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] for vif in network_info: [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return self._sync_wrapper(fn, *args, **kwargs) [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self.wait() [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self[:] = self._gt.wait() [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return self._exit_event.wait() [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] result = hub.switch() [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.736232] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return self.greenlet.switch() [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] result = function(*args, **kwargs) [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] return func(*args, **kwargs) [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] raise e [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] nwinfo = self.network_api.allocate_for_instance( [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] created_port_ids = self._update_ports_for_instance( [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] with excutils.save_and_reraise_exception(): [ 536.736580] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] self.force_reraise() [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] raise self.value [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] updated_port = self._update_port( [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] _ensure_no_port_binding_failure(port) [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] raise exception.PortBindingFailed(port_id=port['id']) [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] nova.exception.PortBindingFailed: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. [ 536.736920] env[61824]: ERROR nova.compute.manager [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] [ 536.742198] env[61824]: DEBUG nova.compute.utils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 536.742645] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.915s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.754035] env[61824]: INFO nova.compute.claims [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 536.767771] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Build of instance 2646b1f8-9785-4517-a1d3-2588cd819f2b was re-scheduled: Binding failed for port 6deea1d5-d0fa-4046-b15f-4d5a06f70d57, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 536.773330] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 536.773330] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Acquiring lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.773330] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Acquired lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.773330] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.893129] env[61824]: DEBUG nova.network.neutron [-] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.937519] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.972784] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.977053] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.977554] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.977554] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.977670] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.978084] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.978293] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.978456] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.978618] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.978776] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.978943] env[61824]: DEBUG nova.virt.hardware [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.980255] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedc3a72-b08a-444b-8fd6-c984763efe49 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.988663] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a08e65-46c7-4252-97fb-757a4906c553 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.112572] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Successfully created port: 35abc3b4-44db-4456-929f-cc09ab5341ff {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.123329] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.123530] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.123823] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.123823] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.124155] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.124155] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.124366] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.124712] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.124897] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.125086] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.125273] env[61824]: DEBUG nova.virt.hardware [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.126191] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e234d1-3eeb-4584-b77d-7f00ec33d7c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.135198] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148a1127-8849-45df-92ed-8874242332b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.154293] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 537.161920] env[61824]: DEBUG oslo.service.loopingcall [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.162220] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 537.162430] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e1ede96-d073-4bcd-ac98-1211003a1227 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.180352] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 537.180352] env[61824]: value = "task-1275312" [ 537.180352] env[61824]: _type = "Task" [ 537.180352] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.188378] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275312, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.199543] env[61824]: DEBUG nova.compute.manager [None req-d74d4211-034a-4b95-a35b-bca977ca9b5a tempest-ServerDiagnosticsV248Test-904838589 tempest-ServerDiagnosticsV248Test-904838589-project-admin] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 537.199997] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a3acd6-0f4c-434d-95f3-3ed030e54043 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.209380] env[61824]: INFO nova.compute.manager [None req-d74d4211-034a-4b95-a35b-bca977ca9b5a tempest-ServerDiagnosticsV248Test-904838589 tempest-ServerDiagnosticsV248Test-904838589-project-admin] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Retrieving diagnostics [ 537.210178] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6db19c6-7f7a-475b-91e4-41b546bb369c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.313772] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.352891] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.396491] env[61824]: INFO nova.compute.manager [-] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Took 1.13 seconds to deallocate network for instance. [ 537.398926] env[61824]: DEBUG nova.compute.claims [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.399120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.523138] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.692262] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275312, 'name': CreateVM_Task, 'duration_secs': 0.322308} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.692658] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 537.692831] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.692921] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.693277] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 537.693531] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f36a3b0a-6db9-43a7-99e0-fa11c81e65f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.698250] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 537.698250] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]529f858e-415b-d824-e778-b982c8a8a33f" [ 537.698250] env[61824]: _type = "Task" [ 537.698250] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.699044] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.709996] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529f858e-415b-d824-e778-b982c8a8a33f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.993709] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322442c0-4499-4777-91dd-618972a9910b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.003589] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc235ab-3a24-48da-8d96-86fdf5eae00f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.036278] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Releasing lock "refresh_cache-2646b1f8-9785-4517-a1d3-2588cd819f2b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.036508] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 538.036682] env[61824]: DEBUG nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.036841] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.039703] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f91eca2-d38f-45be-978a-425138f83e9d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.050807] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcf76b4-71ab-4434-bfcc-e5bf22b05c36 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.063267] env[61824]: DEBUG nova.compute.provider_tree [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.124994] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.204745] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.205649] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.206017] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.206092] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b503bac4-01bc-4804-9b15-511a573ac1fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.216190] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529f858e-415b-d824-e778-b982c8a8a33f, 'name': SearchDatastore_Task, 'duration_secs': 0.030383} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.217400] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.217769] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 538.217865] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.217999] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.218294] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 538.218589] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-790778d7-2768-44a7-ab63-237622170096 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.222986] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec2f239-2253-403a-ac9f-cefe93946d05 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.240571] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 538.240571] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 538.241103] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa9029ec-7907-4f9c-b46c-42c5ddbaf065 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.248544] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8a78dbd9-6ac2-4652-85e7-7ee59be346f4 could not be found. [ 538.248749] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 538.248960] env[61824]: INFO nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 538.249173] env[61824]: DEBUG oslo.service.loopingcall [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.251900] env[61824]: ERROR nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. [ 538.251900] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.251900] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.251900] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.251900] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.251900] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.251900] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.251900] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.251900] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.251900] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 538.251900] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.251900] env[61824]: ERROR nova.compute.manager raise self.value [ 538.251900] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.251900] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.251900] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.251900] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.252361] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.252361] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.252361] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. [ 538.252361] env[61824]: ERROR nova.compute.manager [ 538.252361] env[61824]: Traceback (most recent call last): [ 538.252361] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.252361] env[61824]: listener.cb(fileno) [ 538.252361] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.252361] env[61824]: result = function(*args, **kwargs) [ 538.252361] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.252361] env[61824]: return func(*args, **kwargs) [ 538.252361] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.252361] env[61824]: raise e [ 538.252361] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.252361] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 538.252361] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.252361] env[61824]: created_port_ids = self._update_ports_for_instance( [ 538.252361] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.252361] env[61824]: with excutils.save_and_reraise_exception(): [ 538.252361] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.252361] env[61824]: self.force_reraise() [ 538.252361] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.252361] env[61824]: raise self.value [ 538.252361] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.252361] env[61824]: updated_port = self._update_port( [ 538.252361] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.252361] env[61824]: _ensure_no_port_binding_failure(port) [ 538.252361] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.252361] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.253162] env[61824]: nova.exception.PortBindingFailed: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. [ 538.253162] env[61824]: Removing descriptor: 15 [ 538.253646] env[61824]: DEBUG nova.compute.manager [-] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.253764] env[61824]: DEBUG nova.network.neutron [-] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.258414] env[61824]: ERROR nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Traceback (most recent call last): [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] yield resources [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self.driver.spawn(context, instance, image_meta, [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] vm_ref = self.build_virtual_machine(instance, [ 538.258414] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] for vif in network_info: [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return self._sync_wrapper(fn, *args, **kwargs) [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self.wait() [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self[:] = self._gt.wait() [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return self._exit_event.wait() [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.258889] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] result = hub.switch() [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return self.greenlet.switch() [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] result = function(*args, **kwargs) [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return func(*args, **kwargs) [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] raise e [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] nwinfo = self.network_api.allocate_for_instance( [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] created_port_ids = self._update_ports_for_instance( [ 538.259499] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] with excutils.save_and_reraise_exception(): [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self.force_reraise() [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] raise self.value [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] updated_port = self._update_port( [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] _ensure_no_port_binding_failure(port) [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] raise exception.PortBindingFailed(port_id=port['id']) [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] nova.exception.PortBindingFailed: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. [ 538.259928] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] [ 538.262435] env[61824]: INFO nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Terminating instance [ 538.262435] env[61824]: DEBUG nova.compute.manager [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Received event network-vif-deleted-fb461612-dcac-4cb1-80b6-9413277bacb1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.262435] env[61824]: DEBUG nova.compute.manager [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Received event network-changed-c88594d9-61fa-4595-bb48-f998560bf820 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.262435] env[61824]: DEBUG nova.compute.manager [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Refreshing instance network info cache due to event network-changed-c88594d9-61fa-4595-bb48-f998560bf820. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.262435] env[61824]: DEBUG oslo_concurrency.lockutils [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] Acquiring lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.262645] env[61824]: DEBUG oslo_concurrency.lockutils [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] Acquired lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.262645] env[61824]: DEBUG nova.network.neutron [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Refreshing network info cache for port c88594d9-61fa-4595-bb48-f998560bf820 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 538.262645] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.262645] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquired lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.262645] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 538.266776] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 538.266776] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]524527bb-3c9f-f93a-84e1-b6c8a8c30b4a" [ 538.266776] env[61824]: _type = "Task" [ 538.266776] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.279301] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524527bb-3c9f-f93a-84e1-b6c8a8c30b4a, 'name': SearchDatastore_Task, 'duration_secs': 0.008533} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.280604] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-539ba4ab-2f58-4463-a59a-940c3893be34 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.290647] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 538.290647] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5215c440-72c5-f0d0-c684-bb91ecb885dc" [ 538.290647] env[61824]: _type = "Task" [ 538.290647] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.299509] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5215c440-72c5-f0d0-c684-bb91ecb885dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008139} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.299794] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.299994] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 538.300237] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d4d37fa-6b4a-47a2-8dce-ae792e15babd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.304039] env[61824]: DEBUG nova.network.neutron [-] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.306795] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 538.306795] env[61824]: value = "task-1275313" [ 538.306795] env[61824]: _type = "Task" [ 538.306795] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 538.317359] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275313, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.571703] env[61824]: DEBUG nova.scheduler.client.report [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.587434] env[61824]: DEBUG nova.compute.manager [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Received event network-changed-f7dfe812-b4b9-4bb3-8b72-d398ba14bacc {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.587713] env[61824]: DEBUG nova.compute.manager [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Refreshing instance network info cache due to event network-changed-f7dfe812-b4b9-4bb3-8b72-d398ba14bacc. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.587937] env[61824]: DEBUG oslo_concurrency.lockutils [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] Acquiring lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.595021] env[61824]: DEBUG oslo_concurrency.lockutils [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] Acquired lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.595021] env[61824]: DEBUG nova.network.neutron [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Refreshing network info cache for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 538.628762] env[61824]: DEBUG nova.network.neutron [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.732613] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Acquiring lock "9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.732951] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Lock "9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.795079] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.806966] env[61824]: DEBUG nova.network.neutron [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.811126] env[61824]: DEBUG nova.network.neutron [-] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.824876] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275313, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 538.885403] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.906834] env[61824]: DEBUG nova.network.neutron [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.082841] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.087275] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.090999] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.908s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.091230] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.091383] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 539.091674] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.044s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.106605] env[61824]: INFO nova.compute.claims [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.110546] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ad9649-bf37-494f-81be-1c727663f74d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.132374] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee14d85-aedb-4599-8c66-f2f98ae3a552 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.139248] env[61824]: INFO nova.compute.manager [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] Took 1.10 seconds to deallocate network for instance. [ 539.155569] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8429ec8-d045-400c-8900-a872ce5b9548 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.166695] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f9f5a0-8030-4b3e-a009-fc340daa4700 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.200398] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181548MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 539.200580] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.201665] env[61824]: DEBUG nova.network.neutron [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.321028] env[61824]: INFO nova.compute.manager [-] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Took 1.07 seconds to deallocate network for instance. [ 539.321398] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275313, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519939} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.323724] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 539.323856] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 539.324550] env[61824]: DEBUG nova.compute.claims [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.324550] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.328854] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e27f4340-1340-414a-a86a-c6552d1c5de6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.333517] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 539.333517] env[61824]: value = "task-1275315" [ 539.333517] env[61824]: _type = "Task" [ 539.333517] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.342521] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275315, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.389899] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Releasing lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.390089] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.390303] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 539.390644] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a10d8b3a-a7b3-4542-9205-80f4e2433b1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.399986] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ec82a1-a600-46a2-bb84-9fd855ac4a25 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.412112] env[61824]: DEBUG oslo_concurrency.lockutils [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] Releasing lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.412357] env[61824]: DEBUG nova.compute.manager [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Received event network-vif-deleted-c88594d9-61fa-4595-bb48-f998560bf820 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.416028] env[61824]: DEBUG nova.compute.manager [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Received event network-changed-25203994-eda5-4335-95fc-7fa68d2320d9 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 539.416028] env[61824]: DEBUG nova.compute.manager [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Refreshing instance network info cache due to event network-changed-25203994-eda5-4335-95fc-7fa68d2320d9. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 539.416028] env[61824]: DEBUG oslo_concurrency.lockutils [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] Acquiring lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.416028] env[61824]: DEBUG oslo_concurrency.lockutils [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] Acquired lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.416028] env[61824]: DEBUG nova.network.neutron [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Refreshing network info cache for port 25203994-eda5-4335-95fc-7fa68d2320d9 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 539.428000] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab91f88f-6e87-405f-81ab-bf00be2003a8 could not be found. [ 539.428323] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 539.428536] env[61824]: INFO nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 539.428845] env[61824]: DEBUG oslo.service.loopingcall [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 539.429090] env[61824]: DEBUG nova.compute.manager [-] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.429211] env[61824]: DEBUG nova.network.neutron [-] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 539.553635] env[61824]: DEBUG nova.network.neutron [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.605750] env[61824]: DEBUG nova.network.neutron [-] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.611797] env[61824]: DEBUG nova.compute.utils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.613908] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 539.616608] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 539.684434] env[61824]: DEBUG nova.policy [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f805f2ab5ae40fa9041c5f3b0aee043', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9d15cde819d49bca92b32c418cb7de5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 539.848381] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275315, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061084} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 539.848834] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 539.849719] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6e0288e-65ba-42d2-8026-c7cbf78875ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.875771] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 539.876074] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-946af285-474b-4ca2-ab24-1145e81bfe36 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.898796] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 539.898796] env[61824]: value = "task-1275316" [ 539.898796] env[61824]: _type = "Task" [ 539.898796] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.908573] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275316, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.958171] env[61824]: DEBUG nova.network.neutron [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.058280] env[61824]: DEBUG oslo_concurrency.lockutils [req-f07c4c01-d8e0-4436-947d-ae774135847c req-09badae8-9765-4c34-9599-0f98d58e8d6c service nova] Releasing lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.073297] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Successfully created port: a366f5d1-f97b-470e-84b7-1ecdd68589f0 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.108033] env[61824]: DEBUG nova.network.neutron [-] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.121716] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.191919] env[61824]: INFO nova.scheduler.client.report [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Deleted allocations for instance 2646b1f8-9785-4517-a1d3-2588cd819f2b [ 540.312872] env[61824]: DEBUG nova.network.neutron [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.413272] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275316, 'name': ReconfigVM_Task, 'duration_secs': 0.28722} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.413578] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Reconfigured VM instance instance-00000003 to attach disk [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 540.414200] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-64d668c0-34e6-4111-8e6f-ac2db8660530 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.421261] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 540.421261] env[61824]: value = "task-1275317" [ 540.421261] env[61824]: _type = "Task" [ 540.421261] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.422012] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5798d9f-e58e-416f-9838-33860d629f4b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.436715] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13442c2c-e9a8-402e-a3ee-1a9e328f1e68 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.440874] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275317, 'name': Rename_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.473602] env[61824]: ERROR nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. [ 540.473602] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.473602] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.473602] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.473602] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.473602] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.473602] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.473602] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.473602] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.473602] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 540.473602] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.473602] env[61824]: ERROR nova.compute.manager raise self.value [ 540.473602] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.473602] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.473602] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.473602] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.474249] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.474249] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.474249] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. [ 540.474249] env[61824]: ERROR nova.compute.manager [ 540.474249] env[61824]: Traceback (most recent call last): [ 540.474249] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.474249] env[61824]: listener.cb(fileno) [ 540.474249] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.474249] env[61824]: result = function(*args, **kwargs) [ 540.474249] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.474249] env[61824]: return func(*args, **kwargs) [ 540.474249] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.474249] env[61824]: raise e [ 540.474249] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.474249] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 540.474249] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.474249] env[61824]: created_port_ids = self._update_ports_for_instance( [ 540.474249] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.474249] env[61824]: with excutils.save_and_reraise_exception(): [ 540.474249] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.474249] env[61824]: self.force_reraise() [ 540.474249] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.474249] env[61824]: raise self.value [ 540.474249] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.474249] env[61824]: updated_port = self._update_port( [ 540.474249] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.474249] env[61824]: _ensure_no_port_binding_failure(port) [ 540.474249] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.474249] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.475058] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. [ 540.475058] env[61824]: Removing descriptor: 17 [ 540.475058] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485c2125-4b21-4258-bde7-8e535fe07009 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.481347] env[61824]: ERROR nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Traceback (most recent call last): [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] yield resources [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self.driver.spawn(context, instance, image_meta, [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] vm_ref = self.build_virtual_machine(instance, [ 540.481347] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] for vif in network_info: [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return self._sync_wrapper(fn, *args, **kwargs) [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self.wait() [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self[:] = self._gt.wait() [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return self._exit_event.wait() [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.481704] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] result = hub.switch() [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return self.greenlet.switch() [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] result = function(*args, **kwargs) [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return func(*args, **kwargs) [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] raise e [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] nwinfo = self.network_api.allocate_for_instance( [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] created_port_ids = self._update_ports_for_instance( [ 540.482222] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] with excutils.save_and_reraise_exception(): [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self.force_reraise() [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] raise self.value [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] updated_port = self._update_port( [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] _ensure_no_port_binding_failure(port) [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] raise exception.PortBindingFailed(port_id=port['id']) [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] nova.exception.PortBindingFailed: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. [ 540.482618] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] [ 540.483795] env[61824]: INFO nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Terminating instance [ 540.483795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Acquiring lock "ed4352e2-3321-4a05-8a3f-a02e375ecbfb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.483795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Lock "ed4352e2-3321-4a05-8a3f-a02e375ecbfb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.483795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.483988] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquired lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.483988] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.491032] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a1f5aa-34a9-49db-af37-2cf1c23b6419 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.505819] env[61824]: DEBUG nova.compute.provider_tree [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.610817] env[61824]: INFO nova.compute.manager [-] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Took 1.18 seconds to deallocate network for instance. [ 540.614375] env[61824]: DEBUG nova.compute.claims [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 540.614375] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.704953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9edf5f81-25b4-4d56-ac46-b41e52100048 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667 tempest-FloatingIPsAssociationNegativeTestJSON-1048333667-project-member] Lock "2646b1f8-9785-4517-a1d3-2588cd819f2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.009s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.706307] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "2646b1f8-9785-4517-a1d3-2588cd819f2b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 12.038s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.709614] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 2646b1f8-9785-4517-a1d3-2588cd819f2b] During sync_power_state the instance has a pending task (spawning). Skip. [ 540.709614] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "2646b1f8-9785-4517-a1d3-2588cd819f2b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.815680] env[61824]: DEBUG oslo_concurrency.lockutils [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] Releasing lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.815961] env[61824]: DEBUG nova.compute.manager [req-1159976f-a8db-490d-8e18-beaa212904fe req-0c0021ec-04d1-46f6-9595-03f707f38627 service nova] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Received event network-vif-deleted-25203994-eda5-4335-95fc-7fa68d2320d9 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 540.934820] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275317, 'name': Rename_Task, 'duration_secs': 0.132907} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.935122] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 540.935370] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9ac7930-81c1-4235-90fc-543620ce9185 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.942198] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 540.942198] env[61824]: value = "task-1275318" [ 540.942198] env[61824]: _type = "Task" [ 540.942198] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.952628] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275318, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.009446] env[61824]: DEBUG nova.scheduler.client.report [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.027618] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.133340] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.165075] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.165338] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.165510] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.165704] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.165918] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.166034] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.166244] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.166412] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.166587] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.166759] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.166938] env[61824]: DEBUG nova.virt.hardware [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.168144] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4ebfe7-0ad5-4a66-9be8-8c91d1041bbf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.178053] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3a89a2-d785-4ffb-803f-2ad316313ede {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.208417] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.277870] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.458605] env[61824]: DEBUG oslo_vmware.api [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275318, 'name': PowerOnVM_Task, 'duration_secs': 0.51236} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.458867] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 541.459082] env[61824]: DEBUG nova.compute.manager [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 541.460074] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09fb8e5f-1823-4748-ae3b-82eead0f39c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.520755] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.523016] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 541.525457] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.912s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.527834] env[61824]: INFO nova.compute.claims [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 541.652392] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.653690] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.737999] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.764449] env[61824]: ERROR nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. [ 541.764449] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 541.764449] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.764449] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 541.764449] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.764449] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 541.764449] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.764449] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 541.764449] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.764449] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 541.764449] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.764449] env[61824]: ERROR nova.compute.manager raise self.value [ 541.764449] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.764449] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 541.764449] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.764449] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 541.764954] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.764954] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 541.764954] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. [ 541.764954] env[61824]: ERROR nova.compute.manager [ 541.765364] env[61824]: Traceback (most recent call last): [ 541.765460] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 541.765460] env[61824]: listener.cb(fileno) [ 541.765538] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.765538] env[61824]: result = function(*args, **kwargs) [ 541.765630] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.765630] env[61824]: return func(*args, **kwargs) [ 541.765704] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.765704] env[61824]: raise e [ 541.765775] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.765775] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 541.765845] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.765845] env[61824]: created_port_ids = self._update_ports_for_instance( [ 541.765985] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.765985] env[61824]: with excutils.save_and_reraise_exception(): [ 541.766067] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.766067] env[61824]: self.force_reraise() [ 541.766198] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.766198] env[61824]: raise self.value [ 541.766277] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.766277] env[61824]: updated_port = self._update_port( [ 541.768016] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.768016] env[61824]: _ensure_no_port_binding_failure(port) [ 541.768016] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.768016] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 541.768016] env[61824]: nova.exception.PortBindingFailed: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. [ 541.768016] env[61824]: Removing descriptor: 19 [ 541.768016] env[61824]: ERROR nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. [ 541.768016] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Traceback (most recent call last): [ 541.768016] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 541.768016] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] yield resources [ 541.768016] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.768016] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self.driver.spawn(context, instance, image_meta, [ 541.768016] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 541.768016] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] vm_ref = self.build_virtual_machine(instance, [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] for vif in network_info: [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return self._sync_wrapper(fn, *args, **kwargs) [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self.wait() [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self[:] = self._gt.wait() [ 541.768487] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return self._exit_event.wait() [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] result = hub.switch() [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return self.greenlet.switch() [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] result = function(*args, **kwargs) [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return func(*args, **kwargs) [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] raise e [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.768870] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] nwinfo = self.network_api.allocate_for_instance( [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] created_port_ids = self._update_ports_for_instance( [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] with excutils.save_and_reraise_exception(): [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self.force_reraise() [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] raise self.value [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] updated_port = self._update_port( [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] _ensure_no_port_binding_failure(port) [ 541.769289] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.769743] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] raise exception.PortBindingFailed(port_id=port['id']) [ 541.769743] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] nova.exception.PortBindingFailed: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. [ 541.769743] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] [ 541.769743] env[61824]: INFO nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Terminating instance [ 541.776578] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Acquiring lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.776934] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Acquired lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.776934] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 541.780710] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Releasing lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.781605] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.781605] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 541.781605] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64ac0f72-48b1-46ed-90d9-e25f5beb44fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.799127] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56fd4ac-ddb3-4743-a7b9-39b594b13987 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.824650] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e could not be found. [ 541.824907] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.825135] env[61824]: INFO nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 541.825445] env[61824]: DEBUG oslo.service.loopingcall [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.825620] env[61824]: DEBUG nova.compute.manager [-] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.825694] env[61824]: DEBUG nova.network.neutron [-] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 541.862864] env[61824]: DEBUG nova.network.neutron [-] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.983742] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.026736] env[61824]: DEBUG nova.compute.utils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.028125] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.028295] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 542.128820] env[61824]: DEBUG nova.policy [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16aa10bfe6bd4eb4852cf2ec405010aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fe29f3722a841e6a4337b75e77488a1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.307012] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.365229] env[61824]: DEBUG nova.network.neutron [-] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.491586] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.533147] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 542.818763] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a176312c-b351-4d0a-82c4-0d76877f3f84 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.826534] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc795781-6ab8-409c-8ce2-065c91991277 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.857191] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ced7a53-228a-4c33-bc65-54b22505e44a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.864688] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be93ec3-b44a-4117-bb5f-1bbaf4ae3760 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.869405] env[61824]: INFO nova.compute.manager [-] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Took 1.04 seconds to deallocate network for instance. [ 542.871878] env[61824]: DEBUG nova.compute.claims [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.871878] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.880176] env[61824]: DEBUG nova.compute.provider_tree [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.906277] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Successfully created port: 8eca1ae3-facf-41e8-9c29-dee9929d4af8 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 542.995780] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Releasing lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.995972] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.996432] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 542.996536] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c19fbb8e-49de-4b32-9345-72498e9047d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.007088] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1035aa9-a1f6-4953-b42d-7b015442db3e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.030303] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2fb76ee4-a42b-4de6-9889-b9a89cf16533 could not be found. [ 543.030570] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 543.030758] env[61824]: INFO nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Took 0.03 seconds to destroy the instance on the hypervisor. [ 543.030992] env[61824]: DEBUG oslo.service.loopingcall [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.031217] env[61824]: DEBUG nova.compute.manager [-] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.031305] env[61824]: DEBUG nova.network.neutron [-] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 543.056251] env[61824]: DEBUG nova.network.neutron [-] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.098943] env[61824]: DEBUG nova.compute.manager [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Received event network-changed-aea0ad08-56c4-45b0-87b6-c2b41e67f3b0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.099160] env[61824]: DEBUG nova.compute.manager [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Refreshing instance network info cache due to event network-changed-aea0ad08-56c4-45b0-87b6-c2b41e67f3b0. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.099410] env[61824]: DEBUG oslo_concurrency.lockutils [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] Acquiring lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.099888] env[61824]: DEBUG oslo_concurrency.lockutils [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] Acquired lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.099888] env[61824]: DEBUG nova.network.neutron [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Refreshing network info cache for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 543.170727] env[61824]: DEBUG nova.compute.manager [req-38200b08-733b-4038-beea-17040466247d req-b8416fec-bdc3-4984-bf51-fae3d5045906 service nova] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Received event network-vif-deleted-f7dfe812-b4b9-4bb3-8b72-d398ba14bacc {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.384882] env[61824]: DEBUG nova.scheduler.client.report [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.546310] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 543.559290] env[61824]: DEBUG nova.network.neutron [-] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.577020] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:52:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='2017849872',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1155023997',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.577020] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.577020] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.577249] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.577249] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.577249] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.577249] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.577249] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.577404] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.577404] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.577796] env[61824]: DEBUG nova.virt.hardware [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.579110] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57e8ac6-8b93-4cc3-b89d-23e7b35f5ae4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.591157] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355b3d4c-1e87-486a-a7c3-30fc23293edc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.653732] env[61824]: DEBUG nova.network.neutron [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.854221] env[61824]: DEBUG nova.network.neutron [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.891653] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.892175] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 543.897026] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.399s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.068164] env[61824]: INFO nova.compute.manager [-] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Took 1.04 seconds to deallocate network for instance. [ 544.071170] env[61824]: DEBUG nova.compute.claims [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 544.072126] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.341505] env[61824]: ERROR nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. [ 544.341505] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.341505] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.341505] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.341505] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.341505] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.341505] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.341505] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.341505] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.341505] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 544.341505] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.341505] env[61824]: ERROR nova.compute.manager raise self.value [ 544.341505] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.341505] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.341505] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.341505] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.343051] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.343051] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.343051] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. [ 544.343051] env[61824]: ERROR nova.compute.manager [ 544.343051] env[61824]: Traceback (most recent call last): [ 544.343051] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.343051] env[61824]: listener.cb(fileno) [ 544.343051] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.343051] env[61824]: result = function(*args, **kwargs) [ 544.343051] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.343051] env[61824]: return func(*args, **kwargs) [ 544.343051] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.343051] env[61824]: raise e [ 544.343051] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.343051] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 544.343051] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.343051] env[61824]: created_port_ids = self._update_ports_for_instance( [ 544.343051] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.343051] env[61824]: with excutils.save_and_reraise_exception(): [ 544.343051] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.343051] env[61824]: self.force_reraise() [ 544.343051] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.343051] env[61824]: raise self.value [ 544.343051] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.343051] env[61824]: updated_port = self._update_port( [ 544.343051] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.343051] env[61824]: _ensure_no_port_binding_failure(port) [ 544.343051] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.343051] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.344366] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. [ 544.344366] env[61824]: Removing descriptor: 19 [ 544.344366] env[61824]: ERROR nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Traceback (most recent call last): [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] yield resources [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self.driver.spawn(context, instance, image_meta, [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.344366] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] vm_ref = self.build_virtual_machine(instance, [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] for vif in network_info: [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return self._sync_wrapper(fn, *args, **kwargs) [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self.wait() [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self[:] = self._gt.wait() [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return self._exit_event.wait() [ 544.344803] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] result = hub.switch() [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return self.greenlet.switch() [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] result = function(*args, **kwargs) [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return func(*args, **kwargs) [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] raise e [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] nwinfo = self.network_api.allocate_for_instance( [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.345225] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] created_port_ids = self._update_ports_for_instance( [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] with excutils.save_and_reraise_exception(): [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self.force_reraise() [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] raise self.value [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] updated_port = self._update_port( [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] _ensure_no_port_binding_failure(port) [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.345559] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] raise exception.PortBindingFailed(port_id=port['id']) [ 544.345867] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] nova.exception.PortBindingFailed: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. [ 544.345867] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] [ 544.345867] env[61824]: INFO nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Terminating instance [ 544.348815] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Acquiring lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.349016] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Acquired lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.349192] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.360775] env[61824]: DEBUG oslo_concurrency.lockutils [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] Releasing lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.361034] env[61824]: DEBUG nova.compute.manager [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Received event network-vif-deleted-aea0ad08-56c4-45b0-87b6-c2b41e67f3b0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.361224] env[61824]: DEBUG nova.compute.manager [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Received event network-changed-35abc3b4-44db-4456-929f-cc09ab5341ff {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.361386] env[61824]: DEBUG nova.compute.manager [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Refreshing instance network info cache due to event network-changed-35abc3b4-44db-4456-929f-cc09ab5341ff. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 544.361610] env[61824]: DEBUG oslo_concurrency.lockutils [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] Acquiring lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.361823] env[61824]: DEBUG oslo_concurrency.lockutils [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] Acquired lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.361884] env[61824]: DEBUG nova.network.neutron [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Refreshing network info cache for port 35abc3b4-44db-4456-929f-cc09ab5341ff {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 544.404898] env[61824]: DEBUG nova.compute.utils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.414733] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 544.414733] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 544.480069] env[61824]: DEBUG nova.policy [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '269fa24e7d34476fbdd174f0bf37e3cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0300db3c85124979a9f560b023c1537c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.680129] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1522a1f-cb5e-4c4b-8c8e-4b23aa4fcc52 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.689450] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50b4881-6ebf-42df-8830-55394a890e84 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.729881] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c854fb-1178-423d-81f6-cd7bccc1fb51 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.739435] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde5b686-176d-45c3-ac6c-9ac1478d033b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.756429] env[61824]: DEBUG nova.compute.provider_tree [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.837429] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Successfully created port: 8e9f6b9e-663e-44af-9eef-12955d5ecc17 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.908935] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.911958] env[61824]: DEBUG nova.network.neutron [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.914077] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 545.069234] env[61824]: DEBUG nova.network.neutron [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.162094] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.265924] env[61824]: DEBUG nova.scheduler.client.report [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.278723] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "de5a46be-193a-40d8-aeba-d5d4de24ef95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.279121] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "de5a46be-193a-40d8-aeba-d5d4de24ef95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.361154] env[61824]: INFO nova.compute.manager [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Rebuilding instance [ 545.416724] env[61824]: DEBUG nova.compute.manager [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 545.417784] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bcdcca-f51a-4473-9cce-50986d358609 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.571280] env[61824]: DEBUG oslo_concurrency.lockutils [req-ccb206c5-cdc7-439f-ac49-f4b0287a8227 req-7ca388dc-a8a7-49e0-840c-86fd9987771d service nova] Releasing lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.664890] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Releasing lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.665313] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 545.666253] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 545.666253] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66a9a2c7-1940-4cd5-811b-ac3a58a9bf3d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.676232] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9785f6-a0af-4a56-8bc8-86bb8d8d1656 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.704059] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance abc541cd-4dfc-4071-9b3d-2ec220f225ab could not be found. [ 545.704059] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 545.704238] env[61824]: INFO nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Took 0.04 seconds to destroy the instance on the hypervisor. [ 545.704365] env[61824]: DEBUG oslo.service.loopingcall [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.704853] env[61824]: DEBUG nova.compute.manager [-] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.704853] env[61824]: DEBUG nova.network.neutron [-] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.725277] env[61824]: DEBUG nova.network.neutron [-] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.770511] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.770511] env[61824]: ERROR nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. [ 545.770511] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Traceback (most recent call last): [ 545.770511] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.770511] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self.driver.spawn(context, instance, image_meta, [ 545.770511] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.770511] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.770511] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.770511] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] vm_ref = self.build_virtual_machine(instance, [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] for vif in network_info: [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return self._sync_wrapper(fn, *args, **kwargs) [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self.wait() [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self[:] = self._gt.wait() [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return self._exit_event.wait() [ 545.770803] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] result = hub.switch() [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return self.greenlet.switch() [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] result = function(*args, **kwargs) [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] return func(*args, **kwargs) [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] raise e [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] nwinfo = self.network_api.allocate_for_instance( [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.771285] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] created_port_ids = self._update_ports_for_instance( [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] with excutils.save_and_reraise_exception(): [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] self.force_reraise() [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] raise self.value [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] updated_port = self._update_port( [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] _ensure_no_port_binding_failure(port) [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.771773] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] raise exception.PortBindingFailed(port_id=port['id']) [ 545.772235] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] nova.exception.PortBindingFailed: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. [ 545.772235] env[61824]: ERROR nova.compute.manager [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] [ 545.772235] env[61824]: DEBUG nova.compute.utils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.772235] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.666s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.774623] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Build of instance c9690979-6f9e-4184-80db-b8f9659aa0c2 was re-scheduled: Binding failed for port fb461612-dcac-4cb1-80b6-9413277bacb1, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.776848] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.776848] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Acquiring lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.776848] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Acquired lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.776848] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 545.866953] env[61824]: ERROR nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. [ 545.866953] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.866953] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.866953] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.866953] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.866953] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.866953] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.866953] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.866953] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.866953] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 545.866953] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.866953] env[61824]: ERROR nova.compute.manager raise self.value [ 545.866953] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.866953] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.866953] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.866953] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.867639] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.867639] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.867639] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. [ 545.867639] env[61824]: ERROR nova.compute.manager [ 545.867639] env[61824]: Traceback (most recent call last): [ 545.867639] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.867639] env[61824]: listener.cb(fileno) [ 545.867639] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.867639] env[61824]: result = function(*args, **kwargs) [ 545.867639] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.867639] env[61824]: return func(*args, **kwargs) [ 545.867639] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.867639] env[61824]: raise e [ 545.867639] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.867639] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 545.867639] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.867639] env[61824]: created_port_ids = self._update_ports_for_instance( [ 545.867639] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.867639] env[61824]: with excutils.save_and_reraise_exception(): [ 545.867639] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.867639] env[61824]: self.force_reraise() [ 545.867639] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.867639] env[61824]: raise self.value [ 545.867639] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.867639] env[61824]: updated_port = self._update_port( [ 545.867639] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.867639] env[61824]: _ensure_no_port_binding_failure(port) [ 545.867639] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.867639] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.868383] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. [ 545.868383] env[61824]: Removing descriptor: 19 [ 545.927192] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 545.934133] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 545.934133] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65a9c058-0f3a-4e93-904f-02102101f5ce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.955069] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 545.955069] env[61824]: value = "task-1275322" [ 545.955069] env[61824]: _type = "Task" [ 545.955069] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.970630] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.983327] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.983327] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.983327] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.983830] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.983830] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.983830] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.984289] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.984582] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.984876] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.985873] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.986284] env[61824]: DEBUG nova.virt.hardware [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.987555] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a75c8c3-1ac3-47a6-a378-3c03747bdb46 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.995623] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Acquiring lock "507a2e12-55a1-4d2c-b3d2-577914cc245e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.996118] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Lock "507a2e12-55a1-4d2c-b3d2-577914cc245e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.002517] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964dc646-38be-45b6-8ebf-73fc5651cd95 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.023615] env[61824]: ERROR nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Traceback (most recent call last): [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] yield resources [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self.driver.spawn(context, instance, image_meta, [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] vm_ref = self.build_virtual_machine(instance, [ 546.023615] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] for vif in network_info: [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] return self._sync_wrapper(fn, *args, **kwargs) [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self.wait() [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self[:] = self._gt.wait() [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] return self._exit_event.wait() [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 546.024056] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] current.throw(*self._exc) [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] result = function(*args, **kwargs) [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] return func(*args, **kwargs) [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] raise e [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] nwinfo = self.network_api.allocate_for_instance( [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] created_port_ids = self._update_ports_for_instance( [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] with excutils.save_and_reraise_exception(): [ 546.024379] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self.force_reraise() [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] raise self.value [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] updated_port = self._update_port( [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] _ensure_no_port_binding_failure(port) [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] raise exception.PortBindingFailed(port_id=port['id']) [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] nova.exception.PortBindingFailed: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. [ 546.024714] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] [ 546.024714] env[61824]: INFO nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Terminating instance [ 546.027423] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Acquiring lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.027670] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Acquired lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.027750] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 546.229165] env[61824]: DEBUG nova.network.neutron [-] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.308139] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.410730] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.470313] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275322, 'name': PowerOffVM_Task, 'duration_secs': 0.184717} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.470674] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 546.470829] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 546.471629] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21e93e4d-10d2-4276-b2e8-b0fdc93e7f7f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.481970] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 546.481970] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89bc0902-79dc-48ec-8cb1-3b62864cdbb2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.512108] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 546.512617] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 546.512617] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Deleting the datastore file [datastore2] ff1dac51-f008-4218-bd4c-f0104657a2ee {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.512919] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a9c4fa2-b8a6-48d8-bedd-4fb8116dc279 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.521345] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 546.521345] env[61824]: value = "task-1275324" [ 546.521345] env[61824]: _type = "Task" [ 546.521345] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.543778] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.576322] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b5fc3d-fea7-4505-af23-507444422ace {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.582460] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.591218] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287a8fae-f2d1-4e33-837e-225c512b79fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.625451] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f29daf2-396c-4e62-b2a3-c6f87d43be7d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.633580] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c9d985-4a1f-49bf-8883-b13d471ff9ad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.654600] env[61824]: DEBUG nova.compute.provider_tree [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.724243] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.731952] env[61824]: INFO nova.compute.manager [-] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Took 1.03 seconds to deallocate network for instance. [ 546.734239] env[61824]: DEBUG nova.compute.claims [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.734435] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.916699] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Releasing lock "refresh_cache-c9690979-6f9e-4184-80db-b8f9659aa0c2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.916699] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 546.916699] env[61824]: DEBUG nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.916699] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.944117] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.037670] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101237} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.037928] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 547.038123] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 547.038301] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.066358] env[61824]: DEBUG nova.compute.manager [req-456ad520-47b6-4bcf-8946-13175a9f22e4 req-427521f7-c6f3-4aec-8635-514b43ce2f3f service nova] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Received event network-vif-deleted-35abc3b4-44db-4456-929f-cc09ab5341ff {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.075777] env[61824]: DEBUG nova.compute.manager [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Received event network-changed-a366f5d1-f97b-470e-84b7-1ecdd68589f0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.076024] env[61824]: DEBUG nova.compute.manager [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Refreshing instance network info cache due to event network-changed-a366f5d1-f97b-470e-84b7-1ecdd68589f0. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 547.076219] env[61824]: DEBUG oslo_concurrency.lockutils [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] Acquiring lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.076347] env[61824]: DEBUG oslo_concurrency.lockutils [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] Acquired lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.076815] env[61824]: DEBUG nova.network.neutron [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Refreshing network info cache for port a366f5d1-f97b-470e-84b7-1ecdd68589f0 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 547.161053] env[61824]: DEBUG nova.scheduler.client.report [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.227459] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Releasing lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.227931] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 547.228110] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 547.228952] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dccb0811-cead-422c-8b03-51b517a73ba4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.238824] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0871c12e-da2a-434a-8164-fb4ae7d7a3e0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.272916] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4 could not be found. [ 547.273066] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 547.273281] env[61824]: INFO nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 547.273532] env[61824]: DEBUG oslo.service.loopingcall [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.273760] env[61824]: DEBUG nova.compute.manager [-] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.273849] env[61824]: DEBUG nova.network.neutron [-] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.362891] env[61824]: DEBUG nova.network.neutron [-] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.446843] env[61824]: DEBUG nova.network.neutron [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.613614] env[61824]: DEBUG nova.network.neutron [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.666705] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.895s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.667871] env[61824]: ERROR nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Traceback (most recent call last): [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self.driver.spawn(context, instance, image_meta, [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] vm_ref = self.build_virtual_machine(instance, [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.667871] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] for vif in network_info: [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return self._sync_wrapper(fn, *args, **kwargs) [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self.wait() [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self[:] = self._gt.wait() [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return self._exit_event.wait() [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] result = hub.switch() [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.668259] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return self.greenlet.switch() [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] result = function(*args, **kwargs) [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] return func(*args, **kwargs) [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] raise e [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] nwinfo = self.network_api.allocate_for_instance( [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] created_port_ids = self._update_ports_for_instance( [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] with excutils.save_and_reraise_exception(): [ 547.668651] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] self.force_reraise() [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] raise self.value [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] updated_port = self._update_port( [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] _ensure_no_port_binding_failure(port) [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] raise exception.PortBindingFailed(port_id=port['id']) [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] nova.exception.PortBindingFailed: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. [ 547.668981] env[61824]: ERROR nova.compute.manager [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] [ 547.669283] env[61824]: DEBUG nova.compute.utils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.670970] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.272s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.681285] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Build of instance 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946 was re-scheduled: Binding failed for port c88594d9-61fa-4595-bb48-f998560bf820, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.682437] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.682437] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.682437] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquired lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.682437] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.799331] env[61824]: DEBUG nova.network.neutron [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.866374] env[61824]: DEBUG nova.network.neutron [-] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.952191] env[61824]: INFO nova.compute.manager [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] Took 1.03 seconds to deallocate network for instance. [ 548.078764] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.079291] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.079463] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.079650] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.079795] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.079938] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.080270] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.080651] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.081014] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.081122] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.081364] env[61824]: DEBUG nova.virt.hardware [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.082273] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecdcb4e5-fd11-484c-be3e-90560048cf8e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.091286] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15256106-dce5-472d-959d-5d48ad8550dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.108719] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 548.114830] env[61824]: DEBUG oslo.service.loopingcall [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.115546] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 548.115833] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-495035af-dd0f-4186-bccd-ba95bb2d28f0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.135824] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 548.135824] env[61824]: value = "task-1275326" [ 548.135824] env[61824]: _type = "Task" [ 548.135824] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.144665] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275326, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.215096] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.303905] env[61824]: DEBUG oslo_concurrency.lockutils [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] Releasing lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.305051] env[61824]: DEBUG nova.compute.manager [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Received event network-vif-deleted-a366f5d1-f97b-470e-84b7-1ecdd68589f0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.305298] env[61824]: DEBUG nova.compute.manager [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Received event network-changed-8eca1ae3-facf-41e8-9c29-dee9929d4af8 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 548.305469] env[61824]: DEBUG nova.compute.manager [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Refreshing instance network info cache due to event network-changed-8eca1ae3-facf-41e8-9c29-dee9929d4af8. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.305689] env[61824]: DEBUG oslo_concurrency.lockutils [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] Acquiring lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.305830] env[61824]: DEBUG oslo_concurrency.lockutils [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] Acquired lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.305985] env[61824]: DEBUG nova.network.neutron [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Refreshing network info cache for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 548.347998] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.368255] env[61824]: INFO nova.compute.manager [-] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Took 1.09 seconds to deallocate network for instance. [ 548.374607] env[61824]: DEBUG nova.compute.claims [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.374607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.513858] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c5dff6-1e93-4304-b19d-90bc16351716 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.524925] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f523667-d3fe-48fb-952b-67b4ddf79efd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.562164] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa570fc-80a3-4bcd-b4cc-4192fae59dcf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.573616] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7df7159-f01e-4d85-be2d-d0436a4f09c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.581155] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Acquiring lock "51d54995-0d13-4560-bdd6-7d9f9f5210dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.581494] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Lock "51d54995-0d13-4560-bdd6-7d9f9f5210dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.596932] env[61824]: DEBUG nova.compute.provider_tree [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.647516] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275326, 'name': CreateVM_Task, 'duration_secs': 0.35668} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.650044] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 548.650044] env[61824]: DEBUG oslo_vmware.service [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53f8f71-3682-4c41-b056-146d3a570d73 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.655865] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.656078] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.656447] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 548.657066] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4100667-23a3-49d6-9acf-9549a697278d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.663736] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 548.663736] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52030cdc-d86d-a56b-94d8-bca1e9b41bc2" [ 548.663736] env[61824]: _type = "Task" [ 548.663736] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.681345] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.681528] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 548.681760] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.681901] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.682235] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 548.682517] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b927fd8c-69e5-4017-b37b-886aab85d638 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.709216] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 548.709216] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 548.709216] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f5db64b-e542-45ea-a003-37f428a05041 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.718186] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8983b18-22a1-4bc3-8ae1-31e2cf6aca0c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.725891] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 548.725891] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d9209d-95ef-aeb2-3df1-75e643aeb680" [ 548.725891] env[61824]: _type = "Task" [ 548.725891] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.735574] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d9209d-95ef-aeb2-3df1-75e643aeb680, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.840658] env[61824]: DEBUG nova.network.neutron [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.850199] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Releasing lock "refresh_cache-6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.850450] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 548.850610] env[61824]: DEBUG nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.850773] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.875395] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.992597] env[61824]: INFO nova.scheduler.client.report [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Deleted allocations for instance c9690979-6f9e-4184-80db-b8f9659aa0c2 [ 549.079594] env[61824]: DEBUG nova.compute.manager [None req-965b0d1d-434f-4c30-b05f-886133cf3d0c tempest-ServerDiagnosticsV248Test-904838589 tempest-ServerDiagnosticsV248Test-904838589-project-admin] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 549.081323] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc383ba5-8242-4659-ae09-1a408bec4be6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.089444] env[61824]: INFO nova.compute.manager [None req-965b0d1d-434f-4c30-b05f-886133cf3d0c tempest-ServerDiagnosticsV248Test-904838589 tempest-ServerDiagnosticsV248Test-904838589-project-admin] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Retrieving diagnostics [ 549.090444] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48803b74-f239-4809-b8b1-2a8c1d8839fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.119199] env[61824]: DEBUG nova.scheduler.client.report [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.210305] env[61824]: DEBUG nova.network.neutron [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.237505] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Preparing fetch location {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 549.237896] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Creating directory with path [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 549.238185] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d145ac1-ce04-4f93-99f6-6451a954ddca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.261674] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Created directory with path [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 549.262045] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Fetch image to [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 549.262127] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Downloading image file data 11b2670a-52f3-4a0e-8363-54ca7a7f5264 to [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk on the data store datastore1 {{(pid=61824) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 549.262924] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa9f05d-56c7-4d4a-ba80-f64a2333ed7f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.271508] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ff8713-b66d-4817-b9cb-3bb84508417b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.281437] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5634fc-2c7a-43d1-b930-8449ac405325 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.315034] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53af90d-dea8-40f1-8e7c-cdc1b12f89e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.324020] env[61824]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-ad2d65e0-8959-44ee-a5b3-ff90616b0450 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.344620] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Downloading image file data 11b2670a-52f3-4a0e-8363-54ca7a7f5264 to the data store datastore1 {{(pid=61824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 549.376914] env[61824]: DEBUG nova.network.neutron [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.415234] env[61824]: DEBUG oslo_vmware.rw_handles [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 549.500962] env[61824]: DEBUG oslo_concurrency.lockutils [None req-305fe52d-0de8-4493-8d25-724bb5f11e3e tempest-ServerDiagnosticsTest-1380807476 tempest-ServerDiagnosticsTest-1380807476-project-member] Lock "c9690979-6f9e-4184-80db-b8f9659aa0c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.229s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.502332] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "c9690979-6f9e-4184-80db-b8f9659aa0c2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 20.833s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.502709] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: c9690979-6f9e-4184-80db-b8f9659aa0c2] During sync_power_state the instance has a pending task (spawning). Skip. [ 549.502709] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "c9690979-6f9e-4184-80db-b8f9659aa0c2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.634154] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.636388] env[61824]: ERROR nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Traceback (most recent call last): [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self.driver.spawn(context, instance, image_meta, [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] vm_ref = self.build_virtual_machine(instance, [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.636388] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] for vif in network_info: [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return self._sync_wrapper(fn, *args, **kwargs) [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self.wait() [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self[:] = self._gt.wait() [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return self._exit_event.wait() [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] result = hub.switch() [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.636853] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return self.greenlet.switch() [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] result = function(*args, **kwargs) [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] return func(*args, **kwargs) [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] raise e [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] nwinfo = self.network_api.allocate_for_instance( [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] created_port_ids = self._update_ports_for_instance( [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] with excutils.save_and_reraise_exception(): [ 549.637235] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] self.force_reraise() [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] raise self.value [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] updated_port = self._update_port( [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] _ensure_no_port_binding_failure(port) [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] raise exception.PortBindingFailed(port_id=port['id']) [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] nova.exception.PortBindingFailed: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. [ 549.637607] env[61824]: ERROR nova.compute.manager [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] [ 549.638037] env[61824]: DEBUG nova.compute.utils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 549.638037] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.437s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.642698] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Build of instance be3227f2-3fe1-4767-90a3-55ee57c05df0 was re-scheduled: Binding failed for port 25203994-eda5-4335-95fc-7fa68d2320d9, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 549.643264] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 549.643521] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Acquiring lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.643683] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Acquired lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.643855] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 549.713394] env[61824]: DEBUG oslo_concurrency.lockutils [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] Releasing lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.713707] env[61824]: DEBUG nova.compute.manager [req-d142c83f-9c2d-47e4-85a3-d5ae45afef4f req-a6afece5-8114-4f5b-a15a-50e8bc474a95 service nova] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Received event network-vif-deleted-8eca1ae3-facf-41e8-9c29-dee9929d4af8 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 549.884336] env[61824]: INFO nova.compute.manager [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] Took 1.03 seconds to deallocate network for instance. [ 550.003773] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.100608] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Acquiring lock "e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.101359] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Lock "e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.177624] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.208543] env[61824]: DEBUG oslo_vmware.rw_handles [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Completed reading data from the image iterator. {{(pid=61824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 550.208777] env[61824]: DEBUG oslo_vmware.rw_handles [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 550.275291] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Downloaded image file data 11b2670a-52f3-4a0e-8363-54ca7a7f5264 to vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk on the data store datastore1 {{(pid=61824) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 550.281949] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Caching image {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 550.281949] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Copying Virtual Disk [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk to [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 550.282173] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dd1792d7-c0ed-43da-b6cf-c2aba8bf86d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.292839] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 550.292839] env[61824]: value = "task-1275328" [ 550.292839] env[61824]: _type = "Task" [ 550.292839] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.306035] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275328, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.324994] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.537301] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.676405] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 550.676405] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance ff1dac51-f008-4218-bd4c-f0104657a2ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 550.805255] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275328, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.833663] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Releasing lock "refresh_cache-be3227f2-3fe1-4767-90a3-55ee57c05df0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.834320] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 550.834320] env[61824]: DEBUG nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.834537] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.846413] env[61824]: DEBUG nova.compute.manager [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Received event network-changed-8e9f6b9e-663e-44af-9eef-12955d5ecc17 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.846723] env[61824]: DEBUG nova.compute.manager [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Refreshing instance network info cache due to event network-changed-8e9f6b9e-663e-44af-9eef-12955d5ecc17. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 550.848707] env[61824]: DEBUG oslo_concurrency.lockutils [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] Acquiring lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.849029] env[61824]: DEBUG oslo_concurrency.lockutils [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] Acquired lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.849332] env[61824]: DEBUG nova.network.neutron [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Refreshing network info cache for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 550.861189] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.928682] env[61824]: INFO nova.scheduler.client.report [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Deleted allocations for instance 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946 [ 551.182191] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.304893] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275328, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.944067} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.305218] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Copied Virtual Disk [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk to [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 551.305413] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Deleting the datastore file [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264/tmp-sparse.vmdk {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 551.306869] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec8785af-52fe-4d83-b5ee-18dcdcd0fe30 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.313506] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 551.313506] env[61824]: value = "task-1275329" [ 551.313506] env[61824]: _type = "Task" [ 551.313506] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.322424] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275329, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.368579] env[61824]: DEBUG nova.network.neutron [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.373527] env[61824]: DEBUG nova.network.neutron [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.442566] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24a18edd-dd9c-4388-b44c-dfc9ea456cd1 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.611s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.443666] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 22.774s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.443884] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 6fc17951-d4d8-4ff0-9e93-2f5a15a5e946] During sync_power_state the instance has a pending task (spawning). Skip. [ 551.444072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "6fc17951-d4d8-4ff0-9e93-2f5a15a5e946" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.469108] env[61824]: DEBUG nova.network.neutron [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.686922] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance be3227f2-3fe1-4767-90a3-55ee57c05df0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.687122] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8a78dbd9-6ac2-4652-85e7-7ee59be346f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.687248] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance ab91f88f-6e87-405f-81ab-bf00be2003a8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.687364] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.687478] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 2fb76ee4-a42b-4de6-9889-b9a89cf16533 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.687588] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance abc541cd-4dfc-4071-9b3d-2ec220f225ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.687699] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 551.748046] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.748350] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.748577] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.748761] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.748964] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.755094] env[61824]: INFO nova.compute.manager [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Terminating instance [ 551.755094] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "refresh_cache-90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.755094] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquired lock "refresh_cache-90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.755094] env[61824]: DEBUG nova.network.neutron [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 551.824474] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275329, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.073859} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.824819] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 551.825835] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Moving file from [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad/11b2670a-52f3-4a0e-8363-54ca7a7f5264 to [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264. {{(pid=61824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 551.825835] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-d7430238-c428-4537-bd16-65d3904d8534 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.833417] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 551.833417] env[61824]: value = "task-1275331" [ 551.833417] env[61824]: _type = "Task" [ 551.833417] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.847678] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275331, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.871603] env[61824]: INFO nova.compute.manager [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] Took 1.04 seconds to deallocate network for instance. [ 551.946215] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 551.971587] env[61824]: DEBUG oslo_concurrency.lockutils [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] Releasing lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.971846] env[61824]: DEBUG nova.compute.manager [req-19e436a1-75eb-4c31-817f-e2dcc5c7e7c1 req-413ea84b-df8d-43f8-91e9-6ee858d5acba service nova] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Received event network-vif-deleted-8e9f6b9e-663e-44af-9eef-12955d5ecc17 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.193787] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance a062288b-e785-4dec-98f2-358b54f3d327 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.283481] env[61824]: DEBUG nova.network.neutron [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.348825] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275331, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.029858} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.349090] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] File moved {{(pid=61824) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 552.349320] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Cleaning up location [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 552.349484] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Deleting the datastore file [datastore1] vmware_temp/cbeedf63-07f3-4577-885b-385d76fdd8ad {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 552.349764] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfe87d49-37bc-41f1-88b2-9cc7abfac5bd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.358995] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 552.358995] env[61824]: value = "task-1275332" [ 552.358995] env[61824]: _type = "Task" [ 552.358995] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.370966] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.390097] env[61824]: DEBUG nova.network.neutron [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.477154] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.699829] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.869242] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02842} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.872450] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 552.872450] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d663e433-e63f-4896-bee9-afc92ea6ac94 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.877383] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 552.877383] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a4a9c6-2b0b-fe17-ee70-9a86eb550649" [ 552.877383] env[61824]: _type = "Task" [ 552.877383] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.893608] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a4a9c6-2b0b-fe17-ee70-9a86eb550649, 'name': SearchDatastore_Task, 'duration_secs': 0.00903} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.896869] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Releasing lock "refresh_cache-90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.896869] env[61824]: DEBUG nova.compute.manager [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 552.896869] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.896869] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.897444] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 552.897444] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb636e9d-376d-40f4-991b-60a099a40ae0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.900859] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f476dcdb-4a17-4f52-b3d3-6629ba0d8f9f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.910159] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 552.910159] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc509137-e776-445a-9c9c-fa34436ef4ba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.912762] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 552.912762] env[61824]: value = "task-1275333" [ 552.912762] env[61824]: _type = "Task" [ 552.912762] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.919412] env[61824]: DEBUG oslo_vmware.api [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 552.919412] env[61824]: value = "task-1275334" [ 552.919412] env[61824]: _type = "Task" [ 552.919412] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.927633] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275333, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.928799] env[61824]: INFO nova.scheduler.client.report [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Deleted allocations for instance be3227f2-3fe1-4767-90a3-55ee57c05df0 [ 552.942607] env[61824]: DEBUG oslo_vmware.api [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.202964] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance ed4352e2-3321-4a05-8a3f-a02e375ecbfb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.433762] env[61824]: DEBUG oslo_vmware.api [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275334, 'name': PowerOffVM_Task, 'duration_secs': 0.168553} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.438363] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 553.438363] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 553.438363] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275333, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.438545] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ea863e7-f66d-4526-bf63-384d414a5495 tempest-ImagesOneServerTestJSON-1186675745 tempest-ImagesOneServerTestJSON-1186675745-project-member] Lock "be3227f2-3fe1-4767-90a3-55ee57c05df0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.357s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.439910] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d5d26c54-77da-474d-86e6-13853bbe5497 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.442368] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "be3227f2-3fe1-4767-90a3-55ee57c05df0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 24.773s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.442421] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: be3227f2-3fe1-4767-90a3-55ee57c05df0] During sync_power_state the instance has a pending task (networking). Skip. [ 553.442580] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "be3227f2-3fe1-4767-90a3-55ee57c05df0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.473208] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 553.473444] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 553.473627] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Deleting the datastore file [datastore2] 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 553.474523] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-960b97bc-475a-464d-95b1-eebee351d1c2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.485586] env[61824]: DEBUG oslo_vmware.api [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for the task: (returnval){ [ 553.485586] env[61824]: value = "task-1275336" [ 553.485586] env[61824]: _type = "Task" [ 553.485586] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.495570] env[61824]: DEBUG oslo_vmware.api [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275336, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.707029] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 553.927157] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275333, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.946699] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.995743] env[61824]: DEBUG oslo_vmware.api [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Task: {'id': task-1275336, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.0938} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.996018] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 553.996214] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 553.996486] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 553.996692] env[61824]: INFO nova.compute.manager [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Took 1.10 seconds to destroy the instance on the hypervisor. [ 553.998606] env[61824]: DEBUG oslo.service.loopingcall [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.998606] env[61824]: DEBUG nova.compute.manager [-] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.998606] env[61824]: DEBUG nova.network.neutron [-] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 554.009468] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Acquiring lock "21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.009753] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Lock "21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.065110] env[61824]: DEBUG nova.network.neutron [-] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.213207] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance de5a46be-193a-40d8-aeba-d5d4de24ef95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.426991] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275333, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.482402} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.427423] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 554.427707] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 554.427950] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbb8072d-7f51-44bf-99b8-f665e4be8e6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.435380] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 554.435380] env[61824]: value = "task-1275337" [ 554.435380] env[61824]: _type = "Task" [ 554.435380] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.447162] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275337, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.477592] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.570303] env[61824]: DEBUG nova.network.neutron [-] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.716268] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 507a2e12-55a1-4d2c-b3d2-577914cc245e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.950839] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275337, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066841} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.950839] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 554.954149] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7938f85f-ea39-46ef-823e-982c66a5c17b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.981599] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 554.982178] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83e2cf4a-4bdb-46d1-a942-6a7142c7df1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.008426] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 555.008426] env[61824]: value = "task-1275338" [ 555.008426] env[61824]: _type = "Task" [ 555.008426] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.017976] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275338, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.074426] env[61824]: INFO nova.compute.manager [-] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Took 1.08 seconds to deallocate network for instance. [ 555.225771] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 51d54995-0d13-4560-bdd6-7d9f9f5210dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.518100] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275338, 'name': ReconfigVM_Task, 'duration_secs': 0.270528} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.518429] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Reconfigured VM instance instance-00000003 to attach disk [datastore1] ff1dac51-f008-4218-bd4c-f0104657a2ee/ff1dac51-f008-4218-bd4c-f0104657a2ee.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 555.519161] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b4c1179f-68ce-4055-9eb9-dfdfc8226fb2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.527160] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 555.527160] env[61824]: value = "task-1275339" [ 555.527160] env[61824]: _type = "Task" [ 555.527160] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.537661] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275339, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.585124] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.734942] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.735249] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 555.735412] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 555.817555] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "35c3e1eb-6b98-42f7-8f92-420560feba92" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.817555] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "35c3e1eb-6b98-42f7-8f92-420560feba92" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.043313] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275339, 'name': Rename_Task, 'duration_secs': 0.165613} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.046227] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 556.047028] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5a5375d-6887-4c81-b145-5b2c230b7b73 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.055987] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Waiting for the task: (returnval){ [ 556.055987] env[61824]: value = "task-1275340" [ 556.055987] env[61824]: _type = "Task" [ 556.055987] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.063499] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bca9915-d7fe-461a-b25f-7ddb5c6a03f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.071963] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275340, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.074855] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd627e0-c1a4-47d6-a360-efc4fc7b7095 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.109441] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52594c83-7b54-4d94-8a5b-cdc72fb4e8ee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.117741] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d24b025-99c7-4490-a2f8-507f0e4bc42d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.133317] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 556.575299] env[61824]: DEBUG oslo_vmware.api [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Task: {'id': task-1275340, 'name': PowerOnVM_Task, 'duration_secs': 0.469617} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.580332] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 556.580650] env[61824]: DEBUG nova.compute.manager [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 556.581526] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5beabd1a-0914-4461-a6d3-634a15cdb8f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.636044] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.918343] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Acquiring lock "002a7370-024d-4959-a56d-1a13c260f60e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.918343] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Lock "002a7370-024d-4959-a56d-1a13c260f60e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.106238] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.146506] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 557.146768] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.509s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.148650] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.822s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.155519] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 557.155919] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Getting list of instances from cluster (obj){ [ 557.155919] env[61824]: value = "domain-c8" [ 557.155919] env[61824]: _type = "ClusterComputeResource" [ 557.155919] env[61824]: } {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 557.157222] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b83c1f6-8048-48da-89b6-8df0f90254d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.171464] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Got total of 1 instances {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 557.687120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "46b484d0-1e87-4438-9791-e8199c76d21f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.687120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "46b484d0-1e87-4438-9791-e8199c76d21f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.097067] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6f8b11-3640-4025-9355-ea06e9a0238f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.106308] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce85c91e-36c9-4ddf-af1c-7b6f3a7e9d2e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.148717] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35742512-c63c-42e6-9390-2312ca1377ad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.158026] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df7d685-bef0-41ed-820c-76d453dc2a28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.175663] env[61824]: DEBUG nova.compute.provider_tree [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.275703] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Acquiring lock "95635ec3-cbfd-4dc0-a595-02062a0bda6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.276590] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Lock "95635ec3-cbfd-4dc0-a595-02062a0bda6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.367453] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Acquiring lock "496f00ef-7487-45d9-9356-529f9390544f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.367453] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Lock "496f00ef-7487-45d9-9356-529f9390544f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.682179] env[61824]: DEBUG nova.scheduler.client.report [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.194068] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.046s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.194547] env[61824]: ERROR nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Traceback (most recent call last): [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self.driver.spawn(context, instance, image_meta, [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] vm_ref = self.build_virtual_machine(instance, [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 559.194547] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] for vif in network_info: [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return self._sync_wrapper(fn, *args, **kwargs) [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self.wait() [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self[:] = self._gt.wait() [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return self._exit_event.wait() [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] result = hub.switch() [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 559.194892] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return self.greenlet.switch() [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] result = function(*args, **kwargs) [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] return func(*args, **kwargs) [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] raise e [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] nwinfo = self.network_api.allocate_for_instance( [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] created_port_ids = self._update_ports_for_instance( [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] with excutils.save_and_reraise_exception(): [ 559.195358] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] self.force_reraise() [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] raise self.value [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] updated_port = self._update_port( [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] _ensure_no_port_binding_failure(port) [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] raise exception.PortBindingFailed(port_id=port['id']) [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] nova.exception.PortBindingFailed: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. [ 559.195774] env[61824]: ERROR nova.compute.manager [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] [ 559.196160] env[61824]: DEBUG nova.compute.utils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 559.196698] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.583s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.200706] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Build of instance 8a78dbd9-6ac2-4652-85e7-7ee59be346f4 was re-scheduled: Binding failed for port f7dfe812-b4b9-4bb3-8b72-d398ba14bacc, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 559.200911] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 559.201209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.201291] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.201526] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 559.726989] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.872207] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.136371] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7875c139-9ad1-4f06-a742-48e8b26b30de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.145244] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53b1b99-a23d-4f03-a05b-abdda0a38975 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.157209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.157209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.157209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "ff1dac51-f008-4218-bd4c-f0104657a2ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.157209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.157475] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.157685] env[61824]: INFO nova.compute.manager [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Terminating instance [ 560.159743] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "refresh_cache-ff1dac51-f008-4218-bd4c-f0104657a2ee" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.160117] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquired lock "refresh_cache-ff1dac51-f008-4218-bd4c-f0104657a2ee" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.163188] env[61824]: DEBUG nova.network.neutron [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 560.198027] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae66983-dd4c-454c-863e-e341bcda2f43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.211413] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064d9534-d6a2-4991-95a5-4b16c88e1b3c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.227262] env[61824]: DEBUG nova.compute.provider_tree [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.229594] env[61824]: DEBUG nova.network.neutron [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.308747] env[61824]: DEBUG nova.network.neutron [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.375764] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-8a78dbd9-6ac2-4652-85e7-7ee59be346f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.377868] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 560.381195] env[61824]: DEBUG nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 560.381195] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.402263] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.644821] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Acquiring lock "ad02e0c7-3d12-48ad-a393-26749fd8c502" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.645048] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Lock "ad02e0c7-3d12-48ad-a393-26749fd8c502" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.732719] env[61824]: DEBUG nova.scheduler.client.report [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.816021] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Releasing lock "refresh_cache-ff1dac51-f008-4218-bd4c-f0104657a2ee" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.816021] env[61824]: DEBUG nova.compute.manager [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 560.816021] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 560.816021] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a5a59e0-7055-4441-befa-3094f58f0188 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.825374] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 560.825642] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-713e07bd-ec22-48c0-9ea5-b5fbdbb52fa8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.838753] env[61824]: DEBUG oslo_vmware.api [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 560.838753] env[61824]: value = "task-1275341" [ 560.838753] env[61824]: _type = "Task" [ 560.838753] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.852990] env[61824]: DEBUG oslo_vmware.api [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275341, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.904787] env[61824]: DEBUG nova.network.neutron [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.243383] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.046s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.243383] env[61824]: ERROR nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. [ 561.243383] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Traceback (most recent call last): [ 561.243383] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.243383] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self.driver.spawn(context, instance, image_meta, [ 561.243383] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.243383] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.243383] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.243383] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] vm_ref = self.build_virtual_machine(instance, [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] for vif in network_info: [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return self._sync_wrapper(fn, *args, **kwargs) [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self.wait() [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self[:] = self._gt.wait() [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return self._exit_event.wait() [ 561.244156] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] result = hub.switch() [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return self.greenlet.switch() [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] result = function(*args, **kwargs) [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] return func(*args, **kwargs) [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] raise e [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] nwinfo = self.network_api.allocate_for_instance( [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.244525] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] created_port_ids = self._update_ports_for_instance( [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] with excutils.save_and_reraise_exception(): [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] self.force_reraise() [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] raise self.value [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] updated_port = self._update_port( [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] _ensure_no_port_binding_failure(port) [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.244880] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] raise exception.PortBindingFailed(port_id=port['id']) [ 561.245215] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] nova.exception.PortBindingFailed: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. [ 561.245215] env[61824]: ERROR nova.compute.manager [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] [ 561.245215] env[61824]: DEBUG nova.compute.utils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 561.246952] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.507s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.247460] env[61824]: INFO nova.compute.claims [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 561.251657] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Build of instance ab91f88f-6e87-405f-81ab-bf00be2003a8 was re-scheduled: Binding failed for port aea0ad08-56c4-45b0-87b6-c2b41e67f3b0, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 561.252165] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 561.252396] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.252643] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquired lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.252828] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.351406] env[61824]: DEBUG oslo_vmware.api [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275341, 'name': PowerOffVM_Task, 'duration_secs': 0.157622} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.351598] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 561.351761] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 561.352019] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6a6c4691-6ef6-42ed-b3f9-c487f224fd82 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.386362] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 561.386812] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 561.389019] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Deleting the datastore file [datastore1] ff1dac51-f008-4218-bd4c-f0104657a2ee {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 561.389019] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dbbf15bc-a59e-477c-bd02-1cf849707d2f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.395844] env[61824]: DEBUG oslo_vmware.api [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for the task: (returnval){ [ 561.395844] env[61824]: value = "task-1275343" [ 561.395844] env[61824]: _type = "Task" [ 561.395844] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.411567] env[61824]: INFO nova.compute.manager [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] Took 1.03 seconds to deallocate network for instance. [ 561.413104] env[61824]: DEBUG oslo_vmware.api [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275343, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.799333] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.870624] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "d51370a2-86d5-4c9e-978c-f3318acfd3a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.870624] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "d51370a2-86d5-4c9e-978c-f3318acfd3a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.908171] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "325a641d-2509-49e6-bfbc-0b2cb85c1ff9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.908171] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "325a641d-2509-49e6-bfbc-0b2cb85c1ff9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.913801] env[61824]: DEBUG oslo_vmware.api [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Task: {'id': task-1275343, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109915} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.917370] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 561.917474] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 561.917776] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 561.917855] env[61824]: INFO nova.compute.manager [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Took 1.10 seconds to destroy the instance on the hypervisor. [ 561.918081] env[61824]: DEBUG oslo.service.loopingcall [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 561.920514] env[61824]: DEBUG nova.compute.manager [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.920514] env[61824]: DEBUG nova.network.neutron [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 561.931437] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.950709] env[61824]: DEBUG nova.network.neutron [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.254060] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Acquiring lock "b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.254374] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Lock "b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.431557] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Releasing lock "refresh_cache-ab91f88f-6e87-405f-81ab-bf00be2003a8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.431795] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 562.431974] env[61824]: DEBUG nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.432152] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 562.451563] env[61824]: INFO nova.scheduler.client.report [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted allocations for instance 8a78dbd9-6ac2-4652-85e7-7ee59be346f4 [ 562.458390] env[61824]: DEBUG nova.network.neutron [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.475715] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.738710] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374aeaef-e07e-4322-8da4-ae080e327cb4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.748275] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78dc382b-f4ce-43f2-ae2a-f70007b89026 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.783033] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b87e8d7-5362-4d08-b87c-2f6f5f0a8aba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.792043] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dce7a7e-7c03-4ab2-a0df-887c1b7add63 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.807209] env[61824]: DEBUG nova.compute.provider_tree [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.970231] env[61824]: INFO nova.compute.manager [-] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Took 1.05 seconds to deallocate network for instance. [ 562.970231] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7856ba16-79bd-4bf8-801e-ddc74d191f36 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "8a78dbd9-6ac2-4652-85e7-7ee59be346f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.389s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.970521] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "8a78dbd9-6ac2-4652-85e7-7ee59be346f4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 34.300s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.974474] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da3213c6-3f9d-41e8-aba7-1ee1a81042dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.982637] env[61824]: DEBUG nova.network.neutron [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.995429] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d9cf64-e6f7-472a-b718-b5f2adadad19 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.310936] env[61824]: DEBUG nova.scheduler.client.report [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.475126] env[61824]: DEBUG nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 563.482080] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.485807] env[61824]: INFO nova.compute.manager [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: ab91f88f-6e87-405f-81ab-bf00be2003a8] Took 1.05 seconds to deallocate network for instance. [ 563.541706] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8a78dbd9-6ac2-4652-85e7-7ee59be346f4] During the sync_power process the instance has moved from host None to host cpu-1 [ 563.541706] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "8a78dbd9-6ac2-4652-85e7-7ee59be346f4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.571s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.816580] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.817023] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.820296] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.839s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.820505] env[61824]: DEBUG nova.objects.instance [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 564.021281] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.334680] env[61824]: DEBUG nova.compute.utils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 564.334680] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 564.334680] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 564.404166] env[61824]: DEBUG nova.policy [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c77f85a576dc4792bc7d2599cdd29d24', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '84436c4850f54b70b4156db03cb8b3c5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 564.527755] env[61824]: INFO nova.scheduler.client.report [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Deleted allocations for instance ab91f88f-6e87-405f-81ab-bf00be2003a8 [ 564.837022] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6358eed1-b30f-4418-b0fe-e674a538acee tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.837022] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.964s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.844025] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 565.040738] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2bba5c7d-11b9-4a4d-a86c-2eb9bd917b75 tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "ab91f88f-6e87-405f-81ab-bf00be2003a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.112s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.051997] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Successfully created port: d1dd40f9-1c20-4fd0-ab32-c304f1d02105 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 565.525792] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Acquiring lock "3cd00254-b61b-498a-946b-ff2028230e25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.526457] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Lock "3cd00254-b61b-498a-946b-ff2028230e25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.545988] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.646876] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "e56a6950-f842-45fb-a0a0-0defe3b3c30f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.647291] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "e56a6950-f842-45fb-a0a0-0defe3b3c30f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.770249] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ee7441-38ef-47f3-9183-ce6ac5e44ded {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.779575] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05254a4b-6f0f-4b0e-878e-93c9b447e05f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.810726] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3588827d-af9a-441e-b6c7-2d5fadf08692 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.820398] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83487595-4603-4f95-b23b-2b2296c4e3c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.835594] env[61824]: DEBUG nova.compute.provider_tree [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.854846] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.891224] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.891471] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.891625] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.891803] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.891946] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.892169] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.892394] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.892554] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.892713] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.892873] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.893099] env[61824]: DEBUG nova.virt.hardware [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.893980] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec9703e-7c5d-4fa1-9e8c-40d2c6077432 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.902674] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce56725-d661-4f99-8493-983a2a6be026 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.071374] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.246225] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Acquiring lock "9ce52867-d547-401e-993d-0053eb9da4c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.246742] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Lock "9ce52867-d547-401e-993d-0053eb9da4c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.343043] env[61824]: DEBUG nova.scheduler.client.report [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 566.553666] env[61824]: ERROR nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. [ 566.553666] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 566.553666] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.553666] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 566.553666] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.553666] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 566.553666] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.553666] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 566.553666] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.553666] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 566.553666] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.553666] env[61824]: ERROR nova.compute.manager raise self.value [ 566.553666] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.553666] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 566.553666] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.553666] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 566.554369] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.554369] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 566.554369] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. [ 566.554369] env[61824]: ERROR nova.compute.manager [ 566.554369] env[61824]: Traceback (most recent call last): [ 566.554369] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 566.554369] env[61824]: listener.cb(fileno) [ 566.554369] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.554369] env[61824]: result = function(*args, **kwargs) [ 566.554369] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.554369] env[61824]: return func(*args, **kwargs) [ 566.554369] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.554369] env[61824]: raise e [ 566.554369] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.554369] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 566.554369] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.554369] env[61824]: created_port_ids = self._update_ports_for_instance( [ 566.554369] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.554369] env[61824]: with excutils.save_and_reraise_exception(): [ 566.554369] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.554369] env[61824]: self.force_reraise() [ 566.554369] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.554369] env[61824]: raise self.value [ 566.554369] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.554369] env[61824]: updated_port = self._update_port( [ 566.554369] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.554369] env[61824]: _ensure_no_port_binding_failure(port) [ 566.554369] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.554369] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 566.555248] env[61824]: nova.exception.PortBindingFailed: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. [ 566.555248] env[61824]: Removing descriptor: 17 [ 566.555248] env[61824]: ERROR nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] Traceback (most recent call last): [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] yield resources [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self.driver.spawn(context, instance, image_meta, [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.555248] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] vm_ref = self.build_virtual_machine(instance, [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] for vif in network_info: [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return self._sync_wrapper(fn, *args, **kwargs) [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self.wait() [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self[:] = self._gt.wait() [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return self._exit_event.wait() [ 566.555646] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] result = hub.switch() [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return self.greenlet.switch() [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] result = function(*args, **kwargs) [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return func(*args, **kwargs) [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] raise e [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] nwinfo = self.network_api.allocate_for_instance( [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.556104] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] created_port_ids = self._update_ports_for_instance( [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] with excutils.save_and_reraise_exception(): [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self.force_reraise() [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] raise self.value [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] updated_port = self._update_port( [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] _ensure_no_port_binding_failure(port) [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.556514] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] raise exception.PortBindingFailed(port_id=port['id']) [ 566.556876] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] nova.exception.PortBindingFailed: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. [ 566.556876] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] [ 566.556876] env[61824]: INFO nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Terminating instance [ 566.557983] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Acquiring lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.558155] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Acquired lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.558316] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.653363] env[61824]: DEBUG nova.compute.manager [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Received event network-changed-d1dd40f9-1c20-4fd0-ab32-c304f1d02105 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 566.653363] env[61824]: DEBUG nova.compute.manager [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Refreshing instance network info cache due to event network-changed-d1dd40f9-1c20-4fd0-ab32-c304f1d02105. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 566.653363] env[61824]: DEBUG oslo_concurrency.lockutils [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] Acquiring lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.849547] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.850197] env[61824]: ERROR nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Traceback (most recent call last): [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self.driver.spawn(context, instance, image_meta, [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] vm_ref = self.build_virtual_machine(instance, [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.850197] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] for vif in network_info: [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return self._sync_wrapper(fn, *args, **kwargs) [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self.wait() [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self[:] = self._gt.wait() [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return self._exit_event.wait() [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] result = hub.switch() [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.850575] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return self.greenlet.switch() [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] result = function(*args, **kwargs) [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] return func(*args, **kwargs) [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] raise e [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] nwinfo = self.network_api.allocate_for_instance( [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] created_port_ids = self._update_ports_for_instance( [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] with excutils.save_and_reraise_exception(): [ 566.850947] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] self.force_reraise() [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] raise self.value [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] updated_port = self._update_port( [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] _ensure_no_port_binding_failure(port) [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] raise exception.PortBindingFailed(port_id=port['id']) [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] nova.exception.PortBindingFailed: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. [ 566.851335] env[61824]: ERROR nova.compute.manager [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] [ 566.851647] env[61824]: DEBUG nova.compute.utils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.853534] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.782s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.860141] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Build of instance 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e was re-scheduled: Binding failed for port 35abc3b4-44db-4456-929f-cc09ab5341ff, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 566.860141] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 566.860141] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.860141] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquired lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.860348] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 567.093758] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.224141] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.392587] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.660920] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.729770] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Releasing lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.729770] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 567.729770] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 567.729982] env[61824]: DEBUG oslo_concurrency.lockutils [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] Acquired lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.730111] env[61824]: DEBUG nova.network.neutron [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Refreshing network info cache for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 567.731664] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e522d7b7-5ae6-44e5-be9b-7bbc4f685472 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.742515] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a02bb0-410a-4883-8251-135ec48b7127 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.774630] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a062288b-e785-4dec-98f2-358b54f3d327 could not be found. [ 567.774630] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 567.774630] env[61824]: INFO nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Took 0.04 seconds to destroy the instance on the hypervisor. [ 567.774630] env[61824]: DEBUG oslo.service.loopingcall [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.774630] env[61824]: DEBUG nova.compute.manager [-] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.774630] env[61824]: DEBUG nova.network.neutron [-] [instance: a062288b-e785-4dec-98f2-358b54f3d327] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.802643] env[61824]: DEBUG nova.network.neutron [-] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.808757] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f1928d-8e5e-4b4c-acad-17b52510f518 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.821602] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e28de0-ecd2-4852-911a-c6e684fe8aa8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.857274] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebfa10d-ad74-4294-b51b-ec7c7bf1407a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.865944] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdfcd6f-9b44-4eb1-aa9a-7850619886d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.885942] env[61824]: DEBUG nova.compute.provider_tree [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.163253] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Releasing lock "refresh_cache-4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.163569] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 568.163741] env[61824]: DEBUG nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.163907] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.185518] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.258088] env[61824]: DEBUG nova.network.neutron [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.305062] env[61824]: DEBUG nova.network.neutron [-] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.374225] env[61824]: DEBUG nova.network.neutron [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.390437] env[61824]: DEBUG nova.scheduler.client.report [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.687841] env[61824]: DEBUG nova.network.neutron [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.810326] env[61824]: INFO nova.compute.manager [-] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Took 1.03 seconds to deallocate network for instance. [ 568.812536] env[61824]: DEBUG nova.compute.claims [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 568.812989] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.877655] env[61824]: DEBUG oslo_concurrency.lockutils [req-dc6413a2-b98d-49e5-ba50-1b68d4a52068 req-2ac1a176-e7c8-4f1e-85a7-aabe1fafb085 service nova] Releasing lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.898291] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.898291] env[61824]: ERROR nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. [ 568.898291] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Traceback (most recent call last): [ 568.898291] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.898291] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self.driver.spawn(context, instance, image_meta, [ 568.898291] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.898291] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.898291] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.898291] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] vm_ref = self.build_virtual_machine(instance, [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] for vif in network_info: [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return self._sync_wrapper(fn, *args, **kwargs) [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self.wait() [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self[:] = self._gt.wait() [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return self._exit_event.wait() [ 568.898626] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] result = hub.switch() [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return self.greenlet.switch() [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] result = function(*args, **kwargs) [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] return func(*args, **kwargs) [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] raise e [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] nwinfo = self.network_api.allocate_for_instance( [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.899062] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] created_port_ids = self._update_ports_for_instance( [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] with excutils.save_and_reraise_exception(): [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] self.force_reraise() [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] raise self.value [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] updated_port = self._update_port( [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] _ensure_no_port_binding_failure(port) [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.899491] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] raise exception.PortBindingFailed(port_id=port['id']) [ 568.899891] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] nova.exception.PortBindingFailed: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. [ 568.899891] env[61824]: ERROR nova.compute.manager [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] [ 568.899891] env[61824]: DEBUG nova.compute.utils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 568.903785] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Build of instance 2fb76ee4-a42b-4de6-9889-b9a89cf16533 was re-scheduled: Binding failed for port a366f5d1-f97b-470e-84b7-1ecdd68589f0, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 568.903785] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 568.903785] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Acquiring lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.903785] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Acquired lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.904593] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.904593] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.169s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.911877] env[61824]: DEBUG nova.compute.manager [req-64131ef5-2980-4041-8e2f-75abf304a3f8 req-9a548165-0333-4155-88ed-f952a72e6efb service nova] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Received event network-vif-deleted-d1dd40f9-1c20-4fd0-ab32-c304f1d02105 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 569.193222] env[61824]: INFO nova.compute.manager [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e] Took 1.03 seconds to deallocate network for instance. [ 569.432097] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.573777] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.842376] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4f2456-cdbf-49da-b4c1-41f2dda033b4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.850977] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d0dccc-1c70-4284-a1af-320abca324ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.883552] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89a6acc-94c5-4386-8fd1-85e0cac9d2bb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.892029] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d0f34e-23e4-4baa-96f9-c782d6618896 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.909846] env[61824]: DEBUG nova.compute.provider_tree [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.077321] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Releasing lock "refresh_cache-2fb76ee4-a42b-4de6-9889-b9a89cf16533" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.077680] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 570.077931] env[61824]: DEBUG nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 570.078193] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.099975] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.232029] env[61824]: INFO nova.scheduler.client.report [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Deleted allocations for instance 4d9f7360-f6f1-4f3a-8c47-27b6ba26950e [ 570.413073] env[61824]: DEBUG nova.scheduler.client.report [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.604060] env[61824]: DEBUG nova.network.neutron [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.743130] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48648b0e-7d96-4492-8803-36609cf8523a tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "4d9f7360-f6f1-4f3a-8c47-27b6ba26950e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.594s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.918953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.015s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.919462] env[61824]: ERROR nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Traceback (most recent call last): [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self.driver.spawn(context, instance, image_meta, [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self._vmops.spawn(context, instance, image_meta, injected_files, [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] vm_ref = self.build_virtual_machine(instance, [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] vif_infos = vmwarevif.get_vif_info(self._session, [ 570.919462] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] for vif in network_info: [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return self._sync_wrapper(fn, *args, **kwargs) [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self.wait() [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self[:] = self._gt.wait() [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return self._exit_event.wait() [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] result = hub.switch() [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 570.919892] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return self.greenlet.switch() [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] result = function(*args, **kwargs) [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] return func(*args, **kwargs) [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] raise e [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] nwinfo = self.network_api.allocate_for_instance( [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] created_port_ids = self._update_ports_for_instance( [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] with excutils.save_and_reraise_exception(): [ 570.920339] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] self.force_reraise() [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] raise self.value [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] updated_port = self._update_port( [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] _ensure_no_port_binding_failure(port) [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] raise exception.PortBindingFailed(port_id=port['id']) [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] nova.exception.PortBindingFailed: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. [ 570.920722] env[61824]: ERROR nova.compute.manager [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] [ 570.921046] env[61824]: DEBUG nova.compute.utils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 570.923473] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.549s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.931136] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Build of instance abc541cd-4dfc-4071-9b3d-2ec220f225ab was re-scheduled: Binding failed for port 8eca1ae3-facf-41e8-9c29-dee9929d4af8, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 570.931136] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 570.931136] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Acquiring lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 570.931136] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Acquired lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 570.931613] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 571.106195] env[61824]: INFO nova.compute.manager [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] [instance: 2fb76ee4-a42b-4de6-9889-b9a89cf16533] Took 1.03 seconds to deallocate network for instance. [ 571.246218] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.455069] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.584244] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.779126] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.941383] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc43402-ae71-426b-95e5-738e6a495918 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.952869] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275fb2f1-2ac2-4587-9750-831de01830bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.986507] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98ff86f-1510-4a48-b803-110a7b4b2bdd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.995481] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dd53459-4f7a-4a39-9ea1-dfca36afbd66 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.011813] env[61824]: DEBUG nova.compute.provider_tree [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.088590] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Releasing lock "refresh_cache-abc541cd-4dfc-4071-9b3d-2ec220f225ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 572.088986] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 572.088986] env[61824]: DEBUG nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 572.089231] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 572.139938] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 572.144595] env[61824]: INFO nova.scheduler.client.report [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Deleted allocations for instance 2fb76ee4-a42b-4de6-9889-b9a89cf16533 [ 572.153901] env[61824]: DEBUG nova.network.neutron [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.516406] env[61824]: DEBUG nova.scheduler.client.report [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.656542] env[61824]: INFO nova.compute.manager [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] [instance: abc541cd-4dfc-4071-9b3d-2ec220f225ab] Took 0.57 seconds to deallocate network for instance. [ 572.663092] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9816aa62-113c-4255-95fc-8f271284febe tempest-VolumesAssistedSnapshotsTest-1993469188 tempest-VolumesAssistedSnapshotsTest-1993469188-project-member] Lock "2fb76ee4-a42b-4de6-9889-b9a89cf16533" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.879s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.904687] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "7a93d1c2-20d6-438e-bc42-1f9aa0219f91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.904914] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "7a93d1c2-20d6-438e-bc42-1f9aa0219f91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.023928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.100s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.024602] env[61824]: ERROR nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Traceback (most recent call last): [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self.driver.spawn(context, instance, image_meta, [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] vm_ref = self.build_virtual_machine(instance, [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 573.024602] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] for vif in network_info: [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] return self._sync_wrapper(fn, *args, **kwargs) [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self.wait() [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self[:] = self._gt.wait() [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] return self._exit_event.wait() [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] current.throw(*self._exc) [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 573.025032] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] result = function(*args, **kwargs) [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] return func(*args, **kwargs) [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] raise e [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] nwinfo = self.network_api.allocate_for_instance( [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] created_port_ids = self._update_ports_for_instance( [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] with excutils.save_and_reraise_exception(): [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] self.force_reraise() [ 573.025578] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] raise self.value [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] updated_port = self._update_port( [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] _ensure_no_port_binding_failure(port) [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] raise exception.PortBindingFailed(port_id=port['id']) [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] nova.exception.PortBindingFailed: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. [ 573.026083] env[61824]: ERROR nova.compute.manager [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] [ 573.026083] env[61824]: DEBUG nova.compute.utils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 573.026563] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.489s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.027983] env[61824]: INFO nova.compute.claims [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.030574] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Build of instance e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4 was re-scheduled: Binding failed for port 8e9f6b9e-663e-44af-9eef-12955d5ecc17, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 573.031067] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 573.031223] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Acquiring lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.031368] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Acquired lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.031853] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 573.176022] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 573.282891] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "717eca7a-0287-4424-a302-5737f92724a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.283124] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "717eca7a-0287-4424-a302-5737f92724a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.555865] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 573.636650] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "8aeeb8e9-e1ab-463b-bc16-ca68571882aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.636880] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "8aeeb8e9-e1ab-463b-bc16-ca68571882aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.662470] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.706996] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.709814] env[61824]: INFO nova.scheduler.client.report [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Deleted allocations for instance abc541cd-4dfc-4071-9b3d-2ec220f225ab [ 573.997737] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.997959] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.023260] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "037dd0ba-0d50-4c46-8947-7df3ca19e1c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.023525] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "037dd0ba-0d50-4c46-8947-7df3ca19e1c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.165686] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Releasing lock "refresh_cache-e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.166107] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.166107] env[61824]: DEBUG nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.166271] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 574.203017] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 574.217548] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6362ce3-6b54-4f5e-b16b-6422d5b27883 tempest-ServersWithSpecificFlavorTestJSON-195706646 tempest-ServersWithSpecificFlavorTestJSON-195706646-project-member] Lock "abc541cd-4dfc-4071-9b3d-2ec220f225ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.207s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.396130] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a217f8b-f89b-4a35-9cd7-eedfc39e7e0b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.404534] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07d7cae-f48c-4e93-bfb8-a4561b127e76 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.437496] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bf3287-da00-4987-abda-3387eebce4e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.445942] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b71728-3e3a-461e-8fb8-9bd740b0b8c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.460738] env[61824]: DEBUG nova.compute.provider_tree [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.706264] env[61824]: DEBUG nova.network.neutron [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.719985] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 574.963936] env[61824]: DEBUG nova.scheduler.client.report [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.209879] env[61824]: INFO nova.compute.manager [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] [instance: e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4] Took 1.04 seconds to deallocate network for instance. [ 575.246543] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.470857] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.471499] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.474917] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.998s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.476364] env[61824]: INFO nova.compute.claims [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.980684] env[61824]: DEBUG nova.compute.utils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.985411] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.985411] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 576.069048] env[61824]: DEBUG nova.policy [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ea83c4bf2a84f6a8b8655c6b056cd3b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c71aad0181a44710a5ccab24cf3b04b6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.252335] env[61824]: INFO nova.scheduler.client.report [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Deleted allocations for instance e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4 [ 576.486499] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.687144] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Successfully created port: ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.763845] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae725c68-6930-4021-bbe4-424954c234b9 tempest-FloatingIPsAssociationTestJSON-1943361788 tempest-FloatingIPsAssociationTestJSON-1943361788-project-member] Lock "e615e1ff-7914-4dc8-b8df-1a3a2f64c3f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.237s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.908242] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52f3744-0bc5-4c7d-97d6-55fe72b2b74d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.916923] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe813736-360c-43df-9236-46a978d885b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.950628] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5674b3ca-a72f-432c-9eda-667db3aa6522 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.959635] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf21a63-0965-41da-85b3-47d749139547 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.977862] env[61824]: DEBUG nova.compute.provider_tree [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.264970] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.482416] env[61824]: DEBUG nova.scheduler.client.report [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.501858] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.531395] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.531629] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.531879] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.531957] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.532097] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.532241] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.532441] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.532612] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.532747] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.532907] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.533279] env[61824]: DEBUG nova.virt.hardware [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.534162] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4118b0d-e4a5-490a-9c33-376c20fbfba2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.546638] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69c10a2-200e-4636-882f-0f3e9f18fb3a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.795520] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.990458] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.991046] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.994420] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.517s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.995816] env[61824]: INFO nova.compute.claims [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.507494] env[61824]: DEBUG nova.compute.utils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.511329] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.511505] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 578.646494] env[61824]: DEBUG nova.policy [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34c210845d2c4e3cb4c84673d84b2254', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b87eec8d6bf4d66b6cd9e61ce0e9293', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.015019] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 579.446741] env[61824]: DEBUG nova.compute.manager [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Received event network-changed-ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.447573] env[61824]: DEBUG nova.compute.manager [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Refreshing instance network info cache due to event network-changed-ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.447573] env[61824]: DEBUG oslo_concurrency.lockutils [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] Acquiring lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.447573] env[61824]: DEBUG oslo_concurrency.lockutils [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] Acquired lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.447573] env[61824]: DEBUG nova.network.neutron [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Refreshing network info cache for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 579.453628] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cb2584-6e37-4477-b9a6-cb8cd7fc64f4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.466960] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4221cf-b679-49e8-89b6-7a91c25ffb3d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.503464] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e898f3b-1a69-49f0-87ef-b02cd8fc9368 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.511798] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a84552-e084-4b24-99bc-4a895c5ec9fe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.531609] env[61824]: DEBUG nova.compute.provider_tree [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.570974] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Successfully created port: eb95b184-7f01-4fbd-a154-7cc225b378c1 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.803878] env[61824]: ERROR nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. [ 579.803878] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.803878] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.803878] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.803878] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.803878] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.803878] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.803878] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.803878] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.803878] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 579.803878] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.803878] env[61824]: ERROR nova.compute.manager raise self.value [ 579.803878] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.803878] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.803878] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.803878] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.804429] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.804429] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.804429] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. [ 579.804429] env[61824]: ERROR nova.compute.manager [ 579.804429] env[61824]: Traceback (most recent call last): [ 579.804429] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.804429] env[61824]: listener.cb(fileno) [ 579.804429] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.804429] env[61824]: result = function(*args, **kwargs) [ 579.804429] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.804429] env[61824]: return func(*args, **kwargs) [ 579.804429] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.804429] env[61824]: raise e [ 579.804429] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.804429] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 579.804429] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.804429] env[61824]: created_port_ids = self._update_ports_for_instance( [ 579.804429] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.804429] env[61824]: with excutils.save_and_reraise_exception(): [ 579.804429] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.804429] env[61824]: self.force_reraise() [ 579.804429] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.804429] env[61824]: raise self.value [ 579.804429] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.804429] env[61824]: updated_port = self._update_port( [ 579.804429] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.804429] env[61824]: _ensure_no_port_binding_failure(port) [ 579.804429] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.804429] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.805313] env[61824]: nova.exception.PortBindingFailed: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. [ 579.805313] env[61824]: Removing descriptor: 17 [ 579.805313] env[61824]: ERROR nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Traceback (most recent call last): [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] yield resources [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self.driver.spawn(context, instance, image_meta, [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.805313] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] vm_ref = self.build_virtual_machine(instance, [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] for vif in network_info: [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return self._sync_wrapper(fn, *args, **kwargs) [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self.wait() [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self[:] = self._gt.wait() [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return self._exit_event.wait() [ 579.805801] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] result = hub.switch() [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return self.greenlet.switch() [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] result = function(*args, **kwargs) [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return func(*args, **kwargs) [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] raise e [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] nwinfo = self.network_api.allocate_for_instance( [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.806205] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] created_port_ids = self._update_ports_for_instance( [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] with excutils.save_and_reraise_exception(): [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self.force_reraise() [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] raise self.value [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] updated_port = self._update_port( [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] _ensure_no_port_binding_failure(port) [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.806667] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] raise exception.PortBindingFailed(port_id=port['id']) [ 579.807076] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] nova.exception.PortBindingFailed: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. [ 579.807076] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] [ 579.807076] env[61824]: INFO nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Terminating instance [ 579.810109] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Acquiring lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.034633] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 580.037630] env[61824]: DEBUG nova.scheduler.client.report [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.069137] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.069137] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.069137] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.069317] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.069317] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.069317] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.069317] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.069317] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.069495] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.069495] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.069495] env[61824]: DEBUG nova.virt.hardware [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.069740] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0240aa81-824d-442d-8aad-8b08efa2c177 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.078187] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08606606-71e2-40b4-9ce9-bc453c7f9c76 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.113399] env[61824]: DEBUG nova.network.neutron [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.377205] env[61824]: DEBUG nova.network.neutron [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.545025] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.545025] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.548633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.963s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.552692] env[61824]: DEBUG nova.objects.instance [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lazy-loading 'resources' on Instance uuid 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 580.879679] env[61824]: DEBUG oslo_concurrency.lockutils [req-8dc67c5e-fc8d-49dd-a3c8-75414b872e33 req-9a30e26e-20df-404f-918c-2c2521ffbf31 service nova] Releasing lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.881197] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Acquired lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.881395] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 581.055020] env[61824]: DEBUG nova.compute.utils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.059356] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 581.059542] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 581.137715] env[61824]: DEBUG nova.policy [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1d7f006ab244da48e1bfc38a5859932', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e23c7b193932478096b2fbf49a59fd7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.418951] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.498018] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efffc081-64fd-416f-a6a3-8ad909fe91b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.505181] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44629f4d-a024-4fce-a546-1122b9143719 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.547064] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab3c2dd-df58-488a-81e9-4f4c1848ec75 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.554381] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8f9600-0aa5-4289-bb5f-f30ac1159667 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.567380] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.570272] env[61824]: DEBUG nova.compute.provider_tree [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.651855] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.743308] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Successfully created port: 4c302980-ed9e-4f69-a400-e95a9be397a2 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.900756] env[61824]: DEBUG nova.compute.manager [req-fd04253a-6ad1-4b31-9f54-87d96e4b367a req-b2e5f038-d4b7-455a-8362-468b7d400047 service nova] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Received event network-vif-deleted-ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.078682] env[61824]: DEBUG nova.scheduler.client.report [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.155281] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Releasing lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.155878] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 582.156085] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.156386] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45748742-b57a-41ac-972d-b87efe047da2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.167559] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-530bedaa-1bf4-459b-baa4-487e3c8e38c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.196177] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e could not be found. [ 582.196404] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.196581] env[61824]: INFO nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 582.196823] env[61824]: DEBUG oslo.service.loopingcall [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.197038] env[61824]: DEBUG nova.compute.manager [-] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.197131] env[61824]: DEBUG nova.network.neutron [-] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 582.240814] env[61824]: DEBUG nova.network.neutron [-] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.454309] env[61824]: ERROR nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. [ 582.454309] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.454309] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.454309] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.454309] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.454309] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.454309] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.454309] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.454309] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.454309] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 582.454309] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.454309] env[61824]: ERROR nova.compute.manager raise self.value [ 582.454309] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.454309] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.454309] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.454309] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.455090] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.455090] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.455090] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. [ 582.455090] env[61824]: ERROR nova.compute.manager [ 582.455090] env[61824]: Traceback (most recent call last): [ 582.455090] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.455090] env[61824]: listener.cb(fileno) [ 582.455090] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.455090] env[61824]: result = function(*args, **kwargs) [ 582.455090] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.455090] env[61824]: return func(*args, **kwargs) [ 582.455090] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.455090] env[61824]: raise e [ 582.455090] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.455090] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 582.455090] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.455090] env[61824]: created_port_ids = self._update_ports_for_instance( [ 582.455090] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.455090] env[61824]: with excutils.save_and_reraise_exception(): [ 582.455090] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.455090] env[61824]: self.force_reraise() [ 582.455090] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.455090] env[61824]: raise self.value [ 582.455090] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.455090] env[61824]: updated_port = self._update_port( [ 582.455090] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.455090] env[61824]: _ensure_no_port_binding_failure(port) [ 582.455090] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.455090] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.456754] env[61824]: nova.exception.PortBindingFailed: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. [ 582.456754] env[61824]: Removing descriptor: 19 [ 582.456754] env[61824]: ERROR nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Traceback (most recent call last): [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] yield resources [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self.driver.spawn(context, instance, image_meta, [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.456754] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] vm_ref = self.build_virtual_machine(instance, [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] for vif in network_info: [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return self._sync_wrapper(fn, *args, **kwargs) [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self.wait() [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self[:] = self._gt.wait() [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return self._exit_event.wait() [ 582.457408] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] result = hub.switch() [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return self.greenlet.switch() [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] result = function(*args, **kwargs) [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return func(*args, **kwargs) [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] raise e [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] nwinfo = self.network_api.allocate_for_instance( [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.458014] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] created_port_ids = self._update_ports_for_instance( [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] with excutils.save_and_reraise_exception(): [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self.force_reraise() [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] raise self.value [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] updated_port = self._update_port( [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] _ensure_no_port_binding_failure(port) [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.458467] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] raise exception.PortBindingFailed(port_id=port['id']) [ 582.459281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] nova.exception.PortBindingFailed: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. [ 582.459281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] [ 582.459281] env[61824]: INFO nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Terminating instance [ 582.461074] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Acquiring lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.461074] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Acquired lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.461074] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.584876] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.588286] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.593621] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.485s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.593621] env[61824]: DEBUG nova.objects.instance [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 582.622865] env[61824]: INFO nova.scheduler.client.report [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Deleted allocations for instance 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468 [ 582.634808] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.636171] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.636171] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.636171] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.636302] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.636360] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.636595] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.636754] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.636912] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.637079] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.637243] env[61824]: DEBUG nova.virt.hardware [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.640243] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38417d97-d389-40be-9aba-e10adaff06f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.653732] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fabe99a-2b3b-48c8-9408-82d5a573fa73 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.684597] env[61824]: DEBUG nova.compute.manager [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Received event network-changed-eb95b184-7f01-4fbd-a154-7cc225b378c1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 582.684748] env[61824]: DEBUG nova.compute.manager [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Refreshing instance network info cache due to event network-changed-eb95b184-7f01-4fbd-a154-7cc225b378c1. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 582.684927] env[61824]: DEBUG oslo_concurrency.lockutils [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] Acquiring lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.743375] env[61824]: DEBUG nova.network.neutron [-] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.983671] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.062032] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.150145] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0563fb5d-1d84-41d0-b13d-78fc3d643cc8 tempest-ServerDiagnosticsV248Test-1702625868 tempest-ServerDiagnosticsV248Test-1702625868-project-member] Lock "90a21ec5-7ccf-4d37-9ec9-19ba2bfff468" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.401s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.246831] env[61824]: INFO nova.compute.manager [-] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Took 1.05 seconds to deallocate network for instance. [ 583.249794] env[61824]: DEBUG nova.compute.claims [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.250279] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.569207] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Releasing lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.569663] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.569856] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.570202] env[61824]: DEBUG oslo_concurrency.lockutils [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] Acquired lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.570374] env[61824]: DEBUG nova.network.neutron [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Refreshing network info cache for port eb95b184-7f01-4fbd-a154-7cc225b378c1 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 583.571451] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e90043c-5efd-4258-b0fc-7c0e13ce885e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.587971] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61abc5dd-1910-4d6c-af42-15a731b03917 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.605749] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0809895-816e-4fe3-bec4-1ba01852de3b tempest-ServersAdmin275Test-1366564106 tempest-ServersAdmin275Test-1366564106-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.615267] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.133s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.615537] env[61824]: DEBUG nova.objects.instance [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lazy-loading 'resources' on Instance uuid ff1dac51-f008-4218-bd4c-f0104657a2ee {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 583.617607] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ed4352e2-3321-4a05-8a3f-a02e375ecbfb could not be found. [ 583.617839] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 583.617963] env[61824]: INFO nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 583.618458] env[61824]: DEBUG oslo.service.loopingcall [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.618938] env[61824]: DEBUG nova.compute.manager [-] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.619074] env[61824]: DEBUG nova.network.neutron [-] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.641517] env[61824]: DEBUG nova.network.neutron [-] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.103081] env[61824]: DEBUG nova.network.neutron [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.145714] env[61824]: DEBUG nova.network.neutron [-] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.200566] env[61824]: DEBUG nova.network.neutron [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.541120] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803a073c-f11c-4172-830f-7d032cd003fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.547386] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a27e19-b3ff-4483-b969-d42c92c9d7cb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.590719] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77eb357d-a0e0-4b3f-bafd-70656fd86c10 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.599245] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb71a61-49a8-44f4-ae54-8f036fa9e7f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.613237] env[61824]: DEBUG nova.compute.provider_tree [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.649333] env[61824]: INFO nova.compute.manager [-] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Took 1.03 seconds to deallocate network for instance. [ 584.653894] env[61824]: DEBUG nova.compute.claims [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 584.654105] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.676430] env[61824]: ERROR nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. [ 584.676430] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.676430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.676430] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.676430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.676430] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.676430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.676430] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.676430] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.676430] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 584.676430] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.676430] env[61824]: ERROR nova.compute.manager raise self.value [ 584.676430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.676430] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.676430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.676430] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.677037] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.677037] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.677037] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. [ 584.677037] env[61824]: ERROR nova.compute.manager [ 584.677588] env[61824]: Traceback (most recent call last): [ 584.677956] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.677956] env[61824]: listener.cb(fileno) [ 584.678059] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.678059] env[61824]: result = function(*args, **kwargs) [ 584.678149] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.678149] env[61824]: return func(*args, **kwargs) [ 584.678237] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.678237] env[61824]: raise e [ 584.678327] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.678327] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 584.678411] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.678411] env[61824]: created_port_ids = self._update_ports_for_instance( [ 584.678486] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.678486] env[61824]: with excutils.save_and_reraise_exception(): [ 584.678629] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.678629] env[61824]: self.force_reraise() [ 584.678711] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.678711] env[61824]: raise self.value [ 584.678789] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.678789] env[61824]: updated_port = self._update_port( [ 584.678870] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.678870] env[61824]: _ensure_no_port_binding_failure(port) [ 584.679479] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.679479] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.679479] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. [ 584.679479] env[61824]: Removing descriptor: 17 [ 584.679810] env[61824]: ERROR nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Traceback (most recent call last): [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] yield resources [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self.driver.spawn(context, instance, image_meta, [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] vm_ref = self.build_virtual_machine(instance, [ 584.679810] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] for vif in network_info: [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return self._sync_wrapper(fn, *args, **kwargs) [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self.wait() [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self[:] = self._gt.wait() [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return self._exit_event.wait() [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.680258] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] result = hub.switch() [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return self.greenlet.switch() [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] result = function(*args, **kwargs) [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return func(*args, **kwargs) [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] raise e [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] nwinfo = self.network_api.allocate_for_instance( [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] created_port_ids = self._update_ports_for_instance( [ 584.680698] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] with excutils.save_and_reraise_exception(): [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self.force_reraise() [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] raise self.value [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] updated_port = self._update_port( [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] _ensure_no_port_binding_failure(port) [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] raise exception.PortBindingFailed(port_id=port['id']) [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] nova.exception.PortBindingFailed: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. [ 584.681150] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] [ 584.681603] env[61824]: INFO nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Terminating instance [ 584.682221] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.682319] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.682754] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 584.704040] env[61824]: DEBUG oslo_concurrency.lockutils [req-08866c33-56d7-4e79-80c6-56d1b01d8097 req-d75c2541-5ca9-4b9b-ab3b-a192b1ec7cc8 service nova] Releasing lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.940329] env[61824]: DEBUG nova.compute.manager [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Received event network-changed-4c302980-ed9e-4f69-a400-e95a9be397a2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.941036] env[61824]: DEBUG nova.compute.manager [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Refreshing instance network info cache due to event network-changed-4c302980-ed9e-4f69-a400-e95a9be397a2. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 584.941590] env[61824]: DEBUG oslo_concurrency.lockutils [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] Acquiring lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.116495] env[61824]: DEBUG nova.scheduler.client.report [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 585.220063] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.424411] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.624611] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.009s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.630154] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.609s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.631981] env[61824]: INFO nova.compute.claims [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.639388] env[61824]: DEBUG nova.compute.manager [req-de9145de-d945-4505-b008-cdda7f09b343 req-0815127f-d108-40a3-9ac9-9c8474bfdc40 service nova] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Received event network-vif-deleted-eb95b184-7f01-4fbd-a154-7cc225b378c1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 585.661922] env[61824]: INFO nova.scheduler.client.report [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Deleted allocations for instance ff1dac51-f008-4218-bd4c-f0104657a2ee [ 585.927926] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.927926] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 585.927926] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 585.927926] env[61824]: DEBUG oslo_concurrency.lockutils [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] Acquired lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.927926] env[61824]: DEBUG nova.network.neutron [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Refreshing network info cache for port 4c302980-ed9e-4f69-a400-e95a9be397a2 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 585.928761] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b524f52e-84b8-43c4-95d0-07f387a2a73a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.940651] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20d1013-b450-4018-9b5d-42e71deae27c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.963667] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c could not be found. [ 585.963667] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 585.963667] env[61824]: INFO nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 585.963667] env[61824]: DEBUG oslo.service.loopingcall [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.963944] env[61824]: DEBUG nova.compute.manager [-] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.966214] env[61824]: DEBUG nova.network.neutron [-] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 585.997711] env[61824]: DEBUG nova.network.neutron [-] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.176140] env[61824]: DEBUG oslo_concurrency.lockutils [None req-22237cca-fbf2-4d7a-871b-fd45be66122f tempest-ServersAdmin275Test-621042124 tempest-ServersAdmin275Test-621042124-project-member] Lock "ff1dac51-f008-4218-bd4c-f0104657a2ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.021s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.465129] env[61824]: DEBUG nova.network.neutron [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.501237] env[61824]: DEBUG nova.network.neutron [-] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.611115] env[61824]: DEBUG nova.network.neutron [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.006957] env[61824]: INFO nova.compute.manager [-] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Took 1.04 seconds to deallocate network for instance. [ 587.010427] env[61824]: DEBUG nova.compute.claims [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.010605] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.059078] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366fc34f-2276-46ee-aef7-ffe4cdfed145 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.071928] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3559ff-f102-45cb-bfc5-d1328cbf3c51 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.103019] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa2b496-d47c-47ce-b320-435845469db7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.110447] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c880e1d9-ce95-41b5-984b-79c809528e3b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.115929] env[61824]: DEBUG oslo_concurrency.lockutils [req-00c9dc8e-60bd-4b5a-a7ea-c54036af16bc req-4e32116b-d329-4bbd-89ce-e4cea2c13850 service nova] Releasing lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.126534] env[61824]: DEBUG nova.compute.provider_tree [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 587.465137] env[61824]: DEBUG nova.compute.manager [req-9e6caa9d-e045-496d-b4d5-eb2a14c807e2 req-93ea5ed5-4eb4-44b3-bcbe-24104dc21c2e service nova] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Received event network-vif-deleted-4c302980-ed9e-4f69-a400-e95a9be397a2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.636912] env[61824]: DEBUG nova.scheduler.client.report [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.646512] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Acquiring lock "6e2190ac-d287-4810-9d43-740e62ae7b56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.646891] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Lock "6e2190ac-d287-4810-9d43-740e62ae7b56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.149068] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.149577] env[61824]: DEBUG nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 588.153892] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.085s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.155397] env[61824]: INFO nova.compute.claims [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 588.664759] env[61824]: DEBUG nova.compute.utils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.670028] env[61824]: DEBUG nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Not allocating networking since 'none' was specified. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 589.171900] env[61824]: DEBUG nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 589.609180] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b5ede6-18b6-4ff6-bca0-3f20d0f82dc4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.616983] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793c9029-3d71-4fc5-a952-d2cabd240efe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.653427] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9cf16d-6e44-4890-b058-f526927ca7c7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.663384] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a71954-b671-49a1-934a-2efb99015b2d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.678997] env[61824]: DEBUG nova.compute.provider_tree [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.184825] env[61824]: DEBUG nova.scheduler.client.report [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.192034] env[61824]: DEBUG nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 590.222266] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 590.222266] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 590.222266] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 590.222645] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 590.222645] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 590.222645] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 590.222645] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 590.222645] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 590.222811] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 590.222811] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 590.222811] env[61824]: DEBUG nova.virt.hardware [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 590.227972] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95085ce-4b51-4d05-95d7-b32cfb5a78cb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.233507] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f667777-5652-424d-a136-de1637e10b94 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.248703] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 590.253777] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Creating folder: Project (b57cc4b54f8b4db883dcd96762af4f36). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 590.254435] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cbb19e64-18dd-4a98-80ef-15b5cf37201d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.265640] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Created folder: Project (b57cc4b54f8b4db883dcd96762af4f36) in parent group-v274074. [ 590.265836] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Creating folder: Instances. Parent ref: group-v274087. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 590.266099] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e762ba5-d045-4409-90d9-1cfd515bf74b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.274897] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Created folder: Instances in parent group-v274087. [ 590.275091] env[61824]: DEBUG oslo.service.loopingcall [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.275335] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 590.275471] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5e1c812-db24-4a80-8de0-4b6669fc133b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.292985] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 590.292985] env[61824]: value = "task-1275346" [ 590.292985] env[61824]: _type = "Task" [ 590.292985] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.300925] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275346, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.694882] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.540s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.695281] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 590.699604] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.888s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.807354] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275346, 'name': CreateVM_Task, 'duration_secs': 0.273428} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.807695] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 590.808623] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.808795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.809150] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 590.809539] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8110a58-05a1-4725-841d-dc3ea66c941c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.819168] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 590.819168] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]528f5543-80fc-32a9-7421-94337dd7ae45" [ 590.819168] env[61824]: _type = "Task" [ 590.819168] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.829512] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528f5543-80fc-32a9-7421-94337dd7ae45, 'name': SearchDatastore_Task, 'duration_secs': 0.009283} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.829808] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.830045] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 590.830291] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.830666] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.830666] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 590.834343] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94b06e09-52e7-4912-9be9-227cbd152210 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.843160] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 590.843160] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 590.843160] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73254398-85ce-4647-bc9f-708ed1823eac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.848255] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 590.848255] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52999310-afcb-78b2-aa4d-860d990f233e" [ 590.848255] env[61824]: _type = "Task" [ 590.848255] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.860513] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52999310-afcb-78b2-aa4d-860d990f233e, 'name': SearchDatastore_Task, 'duration_secs': 0.007694} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.861976] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5826d292-bbec-4b58-aba3-8e30a3e1722a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.866274] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 590.866274] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]526687f3-9f28-be08-9572-4efbb85fa7bf" [ 590.866274] env[61824]: _type = "Task" [ 590.866274] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.876827] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526687f3-9f28-be08-9572-4efbb85fa7bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.205347] env[61824]: DEBUG nova.compute.utils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 591.215508] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 591.215787] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 591.338285] env[61824]: DEBUG nova.policy [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '045bd4f49ef24512ba5dbb3b260b285d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1aed0c92ab804afcbffc2dbf8aa53a3f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 591.380305] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526687f3-9f28-be08-9572-4efbb85fa7bf, 'name': SearchDatastore_Task, 'duration_secs': 0.008939} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.383159] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.384191] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] de5a46be-193a-40d8-aeba-d5d4de24ef95/de5a46be-193a-40d8-aeba-d5d4de24ef95.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 591.384191] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90c97225-2a1b-4a70-b983-ac576b09e839 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.392951] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 591.392951] env[61824]: value = "task-1275347" [ 591.392951] env[61824]: _type = "Task" [ 591.392951] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.402391] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275347, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.708482] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.712901] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626bc31f-9389-4d33-83d1-3fc2b13fab95 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.721703] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c879d839-c824-401d-8f72-02ed65dae9ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.756594] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bee0d5-34da-4b1d-ae63-1dcc1c446f67 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.763079] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c471db16-7b6c-41be-baf5-56872730b6b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.778327] env[61824]: DEBUG nova.compute.provider_tree [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.905325] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275347, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47285} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.906214] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] de5a46be-193a-40d8-aeba-d5d4de24ef95/de5a46be-193a-40d8-aeba-d5d4de24ef95.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 591.906214] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 591.906553] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db51a4be-3a89-4b88-a4df-aaf80ec17eca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.914789] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 591.914789] env[61824]: value = "task-1275348" [ 591.914789] env[61824]: _type = "Task" [ 591.914789] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.931266] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275348, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.284198] env[61824]: DEBUG nova.scheduler.client.report [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.429280] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275348, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070783} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.429698] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 592.430898] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676f7546-abb9-4211-a77d-0b34efb323c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.459654] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] de5a46be-193a-40d8-aeba-d5d4de24ef95/de5a46be-193a-40d8-aeba-d5d4de24ef95.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 592.459654] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4948fdda-fb84-439c-8652-da40c66b4743 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.482622] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 592.482622] env[61824]: value = "task-1275349" [ 592.482622] env[61824]: _type = "Task" [ 592.482622] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.489729] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275349, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.491283] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Successfully created port: 9f2d1282-fd12-432d-9101-81ae66dcec8e {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 592.725912] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.750196] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.750467] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.750642] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.750820] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.750955] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.751468] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.751743] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.751924] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.752107] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.752269] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.752440] env[61824]: DEBUG nova.virt.hardware [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.753356] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19ccc11-9538-4be8-91ea-18baa1481caa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.762457] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e27175d-8816-4b00-a72f-c5c7b359562d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.791846] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.092s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.792501] env[61824]: ERROR nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] Traceback (most recent call last): [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self.driver.spawn(context, instance, image_meta, [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] vm_ref = self.build_virtual_machine(instance, [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.792501] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] for vif in network_info: [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return self._sync_wrapper(fn, *args, **kwargs) [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self.wait() [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self[:] = self._gt.wait() [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return self._exit_event.wait() [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] result = hub.switch() [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.792782] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return self.greenlet.switch() [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] result = function(*args, **kwargs) [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] return func(*args, **kwargs) [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] raise e [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] nwinfo = self.network_api.allocate_for_instance( [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] created_port_ids = self._update_ports_for_instance( [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] with excutils.save_and_reraise_exception(): [ 592.793143] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] self.force_reraise() [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] raise self.value [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] updated_port = self._update_port( [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] _ensure_no_port_binding_failure(port) [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] raise exception.PortBindingFailed(port_id=port['id']) [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] nova.exception.PortBindingFailed: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. [ 592.793462] env[61824]: ERROR nova.compute.manager [instance: a062288b-e785-4dec-98f2-358b54f3d327] [ 592.793767] env[61824]: DEBUG nova.compute.utils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.794472] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.016s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.796287] env[61824]: INFO nova.compute.claims [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.802958] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Build of instance a062288b-e785-4dec-98f2-358b54f3d327 was re-scheduled: Binding failed for port d1dd40f9-1c20-4fd0-ab32-c304f1d02105, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.802958] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.802958] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Acquiring lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.802958] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Acquired lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.803210] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.990786] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275349, 'name': ReconfigVM_Task, 'duration_secs': 0.503033} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.991487] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Reconfigured VM instance instance-00000011 to attach disk [datastore2] de5a46be-193a-40d8-aeba-d5d4de24ef95/de5a46be-193a-40d8-aeba-d5d4de24ef95.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 592.992269] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-22ba3055-31ed-4293-978a-f6cd0d80ab32 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.998096] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 592.998096] env[61824]: value = "task-1275350" [ 592.998096] env[61824]: _type = "Task" [ 592.998096] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.007192] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275350, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.334944] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.434869] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.511567] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275350, 'name': Rename_Task, 'duration_secs': 0.151098} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.511567] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 593.511567] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9134b2a-5857-414a-bba2-1b6633bfe83c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.519691] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 593.519691] env[61824]: value = "task-1275351" [ 593.519691] env[61824]: _type = "Task" [ 593.519691] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.527940] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275351, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.794363] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "dc662f50-f111-4d26-b017-3ab719004ac2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 593.794363] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "dc662f50-f111-4d26-b017-3ab719004ac2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.939381] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Releasing lock "refresh_cache-a062288b-e785-4dec-98f2-358b54f3d327" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.939425] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 593.940075] env[61824]: DEBUG nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.940075] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.962834] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.034933] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275351, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.296401] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ca4f2aa-159d-4234-aab6-797a53cb0d6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.304058] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883d2625-01e4-4625-8f95-917793ecf754 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.338176] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47022176-f378-4308-9c0b-6bfaa699e622 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.346340] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341f298e-8c37-4481-9961-fda0eb32f5c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.360599] env[61824]: DEBUG nova.compute.provider_tree [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.467035] env[61824]: DEBUG nova.network.neutron [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.530749] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275351, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.865376] env[61824]: DEBUG nova.scheduler.client.report [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.971905] env[61824]: INFO nova.compute.manager [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] [instance: a062288b-e785-4dec-98f2-358b54f3d327] Took 1.03 seconds to deallocate network for instance. [ 595.030707] env[61824]: DEBUG oslo_vmware.api [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275351, 'name': PowerOnVM_Task, 'duration_secs': 1.190891} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 595.031033] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 595.031773] env[61824]: INFO nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Took 4.84 seconds to spawn the instance on the hypervisor. [ 595.031773] env[61824]: DEBUG nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 595.032853] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a895f11c-36e2-4c48-97ac-12e5eb05ed8d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.129740] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "662e815c-2548-4aed-a928-d98f9e1297e7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.129740] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "662e815c-2548-4aed-a928-d98f9e1297e7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.374013] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.375493] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.379313] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.672s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.381092] env[61824]: INFO nova.compute.claims [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.548512] env[61824]: ERROR nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. [ 595.548512] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.548512] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.548512] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.548512] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.548512] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.548512] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.548512] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.548512] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.548512] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 595.548512] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.548512] env[61824]: ERROR nova.compute.manager raise self.value [ 595.548512] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.548512] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.548512] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.548512] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.548982] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.548982] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.548982] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. [ 595.548982] env[61824]: ERROR nova.compute.manager [ 595.548982] env[61824]: Traceback (most recent call last): [ 595.548982] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.548982] env[61824]: listener.cb(fileno) [ 595.548982] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.548982] env[61824]: result = function(*args, **kwargs) [ 595.548982] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.548982] env[61824]: return func(*args, **kwargs) [ 595.548982] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.548982] env[61824]: raise e [ 595.548982] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.548982] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 595.548982] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.548982] env[61824]: created_port_ids = self._update_ports_for_instance( [ 595.548982] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.548982] env[61824]: with excutils.save_and_reraise_exception(): [ 595.548982] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.548982] env[61824]: self.force_reraise() [ 595.548982] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.548982] env[61824]: raise self.value [ 595.548982] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.548982] env[61824]: updated_port = self._update_port( [ 595.548982] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.548982] env[61824]: _ensure_no_port_binding_failure(port) [ 595.548982] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.548982] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.549692] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. [ 595.549692] env[61824]: Removing descriptor: 17 [ 595.553069] env[61824]: ERROR nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Traceback (most recent call last): [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] yield resources [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self.driver.spawn(context, instance, image_meta, [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] vm_ref = self.build_virtual_machine(instance, [ 595.553069] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] for vif in network_info: [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return self._sync_wrapper(fn, *args, **kwargs) [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self.wait() [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self[:] = self._gt.wait() [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return self._exit_event.wait() [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.553652] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] result = hub.switch() [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return self.greenlet.switch() [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] result = function(*args, **kwargs) [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return func(*args, **kwargs) [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] raise e [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] nwinfo = self.network_api.allocate_for_instance( [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] created_port_ids = self._update_ports_for_instance( [ 595.553919] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] with excutils.save_and_reraise_exception(): [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self.force_reraise() [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] raise self.value [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] updated_port = self._update_port( [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] _ensure_no_port_binding_failure(port) [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] raise exception.PortBindingFailed(port_id=port['id']) [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] nova.exception.PortBindingFailed: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. [ 595.554424] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] [ 595.554770] env[61824]: INFO nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Terminating instance [ 595.558058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Acquiring lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.558058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Acquired lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.558058] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.569681] env[61824]: INFO nova.compute.manager [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Took 31.59 seconds to build instance. [ 595.629241] env[61824]: DEBUG nova.compute.manager [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Received event network-changed-9f2d1282-fd12-432d-9101-81ae66dcec8e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.629347] env[61824]: DEBUG nova.compute.manager [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Refreshing instance network info cache due to event network-changed-9f2d1282-fd12-432d-9101-81ae66dcec8e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 595.629566] env[61824]: DEBUG oslo_concurrency.lockutils [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] Acquiring lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.832291] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "e758c89f-e5e2-42e0-926e-c629cb496a89" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.832523] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "e758c89f-e5e2-42e0-926e-c629cb496a89" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.880515] env[61824]: DEBUG nova.compute.utils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.881995] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.881995] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.933311] env[61824]: DEBUG nova.policy [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b7e8936e3d4499a8915c12f660c5100', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80f82211dc4e44bd9f316da84e161095', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.017090] env[61824]: INFO nova.scheduler.client.report [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Deleted allocations for instance a062288b-e785-4dec-98f2-358b54f3d327 [ 596.077015] env[61824]: DEBUG oslo_concurrency.lockutils [None req-48f1a52d-4629-4ebf-8e24-6429d739cdee tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "de5a46be-193a-40d8-aeba-d5d4de24ef95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.796s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.106806] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.177355] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "204fd4be-4a16-4838-a30d-c78339b3834d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.177901] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "204fd4be-4a16-4838-a30d-c78339b3834d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.225559] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.385637] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.535870] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96fac840-7ed7-4209-8eb1-28b12d973a53 tempest-ServersAdminNegativeTestJSON-1660778577 tempest-ServersAdminNegativeTestJSON-1660778577-project-member] Lock "a062288b-e785-4dec-98f2-358b54f3d327" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.831s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.578557] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.675823] env[61824]: DEBUG nova.compute.manager [None req-162633e4-e09b-4ec8-a405-6bfc3051fe07 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 596.676938] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-107fd24c-74ce-4002-8955-df82c1ec4374 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.729676] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Releasing lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.730144] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.730390] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 596.731324] env[61824]: DEBUG oslo_concurrency.lockutils [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] Acquired lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.731502] env[61824]: DEBUG nova.network.neutron [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Refreshing network info cache for port 9f2d1282-fd12-432d-9101-81ae66dcec8e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 596.732985] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78eb2099-02cc-4b59-a292-3d50c195332b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.748712] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b8b462-aabf-4b4e-a4fe-07e218f1b69e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.767480] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Successfully created port: 1ebad03f-4822-4162-a19e-499ae24a52e7 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.782515] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 507a2e12-55a1-4d2c-b3d2-577914cc245e could not be found. [ 596.782766] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 596.783012] env[61824]: INFO nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 596.783296] env[61824]: DEBUG oslo.service.loopingcall [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.786035] env[61824]: DEBUG nova.compute.manager [-] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.786136] env[61824]: DEBUG nova.network.neutron [-] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.802961] env[61824]: DEBUG nova.network.neutron [-] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.873980] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215282c2-dd20-4891-8a71-86897da65e38 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.884298] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73edc7f-b520-4c81-98bb-1d3fa80a7d0a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.929416] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db84b15a-9d13-4d49-9e9a-ef7c7c540f67 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.939600] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b697b7-7706-4887-a4af-7836664c62ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.954101] env[61824]: DEBUG nova.compute.provider_tree [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.039784] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 597.108965] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.189398] env[61824]: INFO nova.compute.manager [None req-162633e4-e09b-4ec8-a405-6bfc3051fe07 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] instance snapshotting [ 597.190135] env[61824]: DEBUG nova.objects.instance [None req-162633e4-e09b-4ec8-a405-6bfc3051fe07 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lazy-loading 'flavor' on Instance uuid de5a46be-193a-40d8-aeba-d5d4de24ef95 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 597.258520] env[61824]: DEBUG nova.network.neutron [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.305008] env[61824]: DEBUG nova.network.neutron [-] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.368572] env[61824]: DEBUG nova.network.neutron [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.437113] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.459057] env[61824]: DEBUG nova.scheduler.client.report [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.463915] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "de5a46be-193a-40d8-aeba-d5d4de24ef95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.466501] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "de5a46be-193a-40d8-aeba-d5d4de24ef95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.466501] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "de5a46be-193a-40d8-aeba-d5d4de24ef95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.466501] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "de5a46be-193a-40d8-aeba-d5d4de24ef95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.466501] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "de5a46be-193a-40d8-aeba-d5d4de24ef95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.468891] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.469125] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.469285] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.469470] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.469613] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.469756] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.469960] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.471503] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.471503] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.471503] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.471503] env[61824]: DEBUG nova.virt.hardware [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.472103] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b09dc0-772e-45bb-9b45-b401789e7378 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.475221] env[61824]: INFO nova.compute.manager [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Terminating instance [ 597.477718] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "refresh_cache-de5a46be-193a-40d8-aeba-d5d4de24ef95" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.477718] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquired lock "refresh_cache-de5a46be-193a-40d8-aeba-d5d4de24ef95" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.477718] env[61824]: DEBUG nova.network.neutron [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.486914] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd92cb1d-2e9f-4512-b963-0fddb020bb8b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.557702] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.699839] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a98c4b4-3e00-4b49-b613-86aa77b01757 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.719721] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88dba6f4-cdde-4d3e-8567-9c135731b420 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.808107] env[61824]: INFO nova.compute.manager [-] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Took 1.02 seconds to deallocate network for instance. [ 597.812691] env[61824]: DEBUG nova.compute.claims [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.812691] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.870468] env[61824]: DEBUG oslo_concurrency.lockutils [req-03a1f8c4-3968-49e8-b81a-e1d449af2f0f req-d86c1ac3-a08e-49ff-bc75-e58a8140bdd8 service nova] Releasing lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.981104] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.600s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.981104] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 597.986039] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.739s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.986678] env[61824]: INFO nova.compute.claims [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 598.015934] env[61824]: DEBUG nova.network.neutron [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.075523] env[61824]: DEBUG nova.network.neutron [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.224902] env[61824]: DEBUG nova.compute.manager [req-162b6d40-a0f8-42e5-ae33-0b6ad79c7cea req-3704ee04-8c3d-48bc-bb41-f2c11fda0f6a service nova] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Received event network-vif-deleted-9f2d1282-fd12-432d-9101-81ae66dcec8e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.230210] env[61824]: DEBUG nova.compute.manager [None req-162633e4-e09b-4ec8-a405-6bfc3051fe07 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Instance disappeared during snapshot {{(pid=61824) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 598.331692] env[61824]: ERROR nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. [ 598.331692] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.331692] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.331692] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.331692] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.331692] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.331692] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.331692] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.331692] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.331692] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 598.331692] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.331692] env[61824]: ERROR nova.compute.manager raise self.value [ 598.331692] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.331692] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.331692] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.331692] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.332104] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.332104] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.332104] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. [ 598.332104] env[61824]: ERROR nova.compute.manager [ 598.332104] env[61824]: Traceback (most recent call last): [ 598.332104] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.332104] env[61824]: listener.cb(fileno) [ 598.332104] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.332104] env[61824]: result = function(*args, **kwargs) [ 598.332104] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.332104] env[61824]: return func(*args, **kwargs) [ 598.332104] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.332104] env[61824]: raise e [ 598.332104] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.332104] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 598.332104] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.332104] env[61824]: created_port_ids = self._update_ports_for_instance( [ 598.332104] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.332104] env[61824]: with excutils.save_and_reraise_exception(): [ 598.332104] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.332104] env[61824]: self.force_reraise() [ 598.332104] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.332104] env[61824]: raise self.value [ 598.332104] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.332104] env[61824]: updated_port = self._update_port( [ 598.332104] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.332104] env[61824]: _ensure_no_port_binding_failure(port) [ 598.332104] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.332104] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.332795] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. [ 598.332795] env[61824]: Removing descriptor: 17 [ 598.332795] env[61824]: ERROR nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Traceback (most recent call last): [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] yield resources [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self.driver.spawn(context, instance, image_meta, [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.332795] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] vm_ref = self.build_virtual_machine(instance, [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] for vif in network_info: [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return self._sync_wrapper(fn, *args, **kwargs) [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self.wait() [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self[:] = self._gt.wait() [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return self._exit_event.wait() [ 598.333083] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] result = hub.switch() [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return self.greenlet.switch() [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] result = function(*args, **kwargs) [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return func(*args, **kwargs) [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] raise e [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] nwinfo = self.network_api.allocate_for_instance( [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.333425] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] created_port_ids = self._update_ports_for_instance( [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] with excutils.save_and_reraise_exception(): [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self.force_reraise() [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] raise self.value [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] updated_port = self._update_port( [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] _ensure_no_port_binding_failure(port) [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.333752] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] raise exception.PortBindingFailed(port_id=port['id']) [ 598.334270] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] nova.exception.PortBindingFailed: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. [ 598.334270] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] [ 598.334270] env[61824]: INFO nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Terminating instance [ 598.337194] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Acquiring lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.337194] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Acquired lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.337194] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.389488] env[61824]: DEBUG nova.compute.manager [None req-162633e4-e09b-4ec8-a405-6bfc3051fe07 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Found 0 images (rotation: 2) {{(pid=61824) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 598.394360] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "f525feb8-d24a-4f89-b640-471163d181ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.394581] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "f525feb8-d24a-4f89-b640-471163d181ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.489093] env[61824]: DEBUG nova.compute.utils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.489093] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 598.489093] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 598.577183] env[61824]: DEBUG nova.policy [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1446fc0b7a0149de93f9b65aadab8402', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45fe7ab6252d4cf3a9b6819a7db08d7f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.580381] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Releasing lock "refresh_cache-de5a46be-193a-40d8-aeba-d5d4de24ef95" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.580765] env[61824]: DEBUG nova.compute.manager [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.580954] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.581824] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b9da39-4f6e-477c-920e-31814699d19e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.595843] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 598.596112] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-796f23fb-96b6-4e2a-9461-013e3ee74d67 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.602850] env[61824]: DEBUG oslo_vmware.api [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 598.602850] env[61824]: value = "task-1275352" [ 598.602850] env[61824]: _type = "Task" [ 598.602850] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.611540] env[61824]: DEBUG oslo_vmware.api [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275352, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.858293] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.939848] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.992894] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.056556] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "9942dd4b-8c04-40ac-b9f4-318b13902535" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.058182] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "9942dd4b-8c04-40ac-b9f4-318b13902535" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.101480] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Successfully created port: fc168f10-6129-41b7-a5e2-5b3c50580726 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 599.114890] env[61824]: DEBUG oslo_vmware.api [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275352, 'name': PowerOffVM_Task, 'duration_secs': 0.118733} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.115188] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 599.115356] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 599.115632] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1cf40fcb-23d5-41e5-a98f-525ce5869a45 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.140652] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 599.140898] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 599.141059] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Deleting the datastore file [datastore2] de5a46be-193a-40d8-aeba-d5d4de24ef95 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 599.141313] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d37d18d0-b860-406d-a53f-41643e09609f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.148366] env[61824]: DEBUG oslo_vmware.api [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for the task: (returnval){ [ 599.148366] env[61824]: value = "task-1275354" [ 599.148366] env[61824]: _type = "Task" [ 599.148366] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.160578] env[61824]: DEBUG oslo_vmware.api [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275354, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.442721] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Releasing lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.443185] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.443338] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.443777] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3e1d4974-69b6-402b-8da9-972fa135267f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.455053] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1e1cf9-1080-40cb-865a-ceeef1d087c5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.467069] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963ca387-5308-4191-9484-99497e8f6548 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.474374] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11f9aae-4fdd-4982-a6fb-c9702a7292f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.481663] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 51d54995-0d13-4560-bdd6-7d9f9f5210dd could not be found. [ 599.481663] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.481886] env[61824]: INFO nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.482158] env[61824]: DEBUG oslo.service.loopingcall [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.482870] env[61824]: DEBUG nova.compute.manager [-] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.482983] env[61824]: DEBUG nova.network.neutron [-] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.515912] env[61824]: DEBUG nova.network.neutron [-] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.517456] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37e9f46-d338-4c06-a10e-a86e1bbfe6b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.530013] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1827aea9-db98-4b4a-8a02-62ea5082ec2f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.544604] env[61824]: DEBUG nova.compute.provider_tree [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.659693] env[61824]: DEBUG oslo_vmware.api [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Task: {'id': task-1275354, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106287} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.659934] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 599.660225] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 599.660417] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.660597] env[61824]: INFO nova.compute.manager [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Took 1.08 seconds to destroy the instance on the hypervisor. [ 599.660829] env[61824]: DEBUG oslo.service.loopingcall [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.661029] env[61824]: DEBUG nova.compute.manager [-] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.661124] env[61824]: DEBUG nova.network.neutron [-] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.680222] env[61824]: DEBUG nova.network.neutron [-] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.024602] env[61824]: DEBUG nova.network.neutron [-] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.026979] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.048417] env[61824]: DEBUG nova.scheduler.client.report [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.060438] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.060688] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.060841] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.061036] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.061188] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.061330] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.061529] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.061686] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.061849] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.062313] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.063244] env[61824]: DEBUG nova.virt.hardware [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.063244] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd7c7ff-91b1-48d4-8aeb-de03660176c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.074934] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488618dd-1add-4b03-b610-77f24a70073b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.182422] env[61824]: DEBUG nova.network.neutron [-] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.298834] env[61824]: DEBUG nova.compute.manager [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Received event network-changed-1ebad03f-4822-4162-a19e-499ae24a52e7 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.299099] env[61824]: DEBUG nova.compute.manager [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Refreshing instance network info cache due to event network-changed-1ebad03f-4822-4162-a19e-499ae24a52e7. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.299352] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] Acquiring lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.299526] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] Acquired lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.299723] env[61824]: DEBUG nova.network.neutron [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Refreshing network info cache for port 1ebad03f-4822-4162-a19e-499ae24a52e7 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 600.379158] env[61824]: ERROR nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. [ 600.379158] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.379158] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.379158] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.379158] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.379158] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.379158] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.379158] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.379158] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.379158] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 600.379158] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.379158] env[61824]: ERROR nova.compute.manager raise self.value [ 600.379158] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.379158] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.379158] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.379158] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.379508] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.379508] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.379508] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. [ 600.379508] env[61824]: ERROR nova.compute.manager [ 600.379508] env[61824]: Traceback (most recent call last): [ 600.379508] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.379508] env[61824]: listener.cb(fileno) [ 600.379508] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.379508] env[61824]: result = function(*args, **kwargs) [ 600.379508] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.379508] env[61824]: return func(*args, **kwargs) [ 600.379508] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.379508] env[61824]: raise e [ 600.379508] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.379508] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 600.379508] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.379508] env[61824]: created_port_ids = self._update_ports_for_instance( [ 600.379508] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.379508] env[61824]: with excutils.save_and_reraise_exception(): [ 600.379508] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.379508] env[61824]: self.force_reraise() [ 600.379508] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.379508] env[61824]: raise self.value [ 600.379508] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.379508] env[61824]: updated_port = self._update_port( [ 600.379508] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.379508] env[61824]: _ensure_no_port_binding_failure(port) [ 600.379508] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.379508] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.380096] env[61824]: nova.exception.PortBindingFailed: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. [ 600.380096] env[61824]: Removing descriptor: 17 [ 600.380096] env[61824]: ERROR nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Traceback (most recent call last): [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] yield resources [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self.driver.spawn(context, instance, image_meta, [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.380096] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] vm_ref = self.build_virtual_machine(instance, [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] for vif in network_info: [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return self._sync_wrapper(fn, *args, **kwargs) [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self.wait() [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self[:] = self._gt.wait() [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return self._exit_event.wait() [ 600.380364] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] result = hub.switch() [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return self.greenlet.switch() [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] result = function(*args, **kwargs) [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return func(*args, **kwargs) [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] raise e [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] nwinfo = self.network_api.allocate_for_instance( [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.380680] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] created_port_ids = self._update_ports_for_instance( [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] with excutils.save_and_reraise_exception(): [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self.force_reraise() [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] raise self.value [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] updated_port = self._update_port( [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] _ensure_no_port_binding_failure(port) [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.381067] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] raise exception.PortBindingFailed(port_id=port['id']) [ 600.381401] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] nova.exception.PortBindingFailed: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. [ 600.381401] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] [ 600.381401] env[61824]: INFO nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Terminating instance [ 600.382369] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Acquiring lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.382535] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Acquired lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.382692] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.529719] env[61824]: INFO nova.compute.manager [-] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Took 1.05 seconds to deallocate network for instance. [ 600.532044] env[61824]: DEBUG nova.compute.claims [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.532224] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.555721] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.556271] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.558768] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.764s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.560245] env[61824]: INFO nova.compute.claims [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.685707] env[61824]: INFO nova.compute.manager [-] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Took 1.02 seconds to deallocate network for instance. [ 600.827475] env[61824]: DEBUG nova.network.neutron [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.906418] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.949352] env[61824]: DEBUG nova.network.neutron [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.010509] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.066754] env[61824]: DEBUG nova.compute.utils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 601.071155] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 601.071387] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 601.157385] env[61824]: DEBUG nova.policy [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb4b3246586f40909ba8b575d1a88623', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '394cc6f362484de7b741b8a2e16b5a4d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.193046] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.455170] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] Releasing lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.455454] env[61824]: DEBUG nova.compute.manager [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Received event network-vif-deleted-1ebad03f-4822-4162-a19e-499ae24a52e7 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.455688] env[61824]: DEBUG nova.compute.manager [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Received event network-changed-fc168f10-6129-41b7-a5e2-5b3c50580726 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 601.455848] env[61824]: DEBUG nova.compute.manager [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Refreshing instance network info cache due to event network-changed-fc168f10-6129-41b7-a5e2-5b3c50580726. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 601.456051] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] Acquiring lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.513774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Releasing lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.513774] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.513774] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 601.513909] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] Acquired lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.514047] env[61824]: DEBUG nova.network.neutron [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Refreshing network info cache for port fc168f10-6129-41b7-a5e2-5b3c50580726 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 601.515117] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91e07391-b879-402e-81c1-0e9dc1c1175d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.529860] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6be6384-1466-4f8e-a679-980d4fd40822 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.553011] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb could not be found. [ 601.553249] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 601.553423] env[61824]: INFO nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 601.553659] env[61824]: DEBUG oslo.service.loopingcall [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.553875] env[61824]: DEBUG nova.compute.manager [-] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.553967] env[61824]: DEBUG nova.network.neutron [-] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.568534] env[61824]: DEBUG nova.network.neutron [-] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.571566] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.582220] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Successfully created port: 2e470afc-43b7-4ca4-959e-f759868f109f {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.045090] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9136da71-b8de-4a52-ab5c-d967657ac3f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.052961] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8093605-03dd-41b6-b155-7c8df5b9c478 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.084709] env[61824]: DEBUG nova.network.neutron [-] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.089614] env[61824]: DEBUG nova.network.neutron [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.092117] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b328c6ec-672c-4717-9c35-080d529e60ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.100529] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c894192-6b27-4ee5-bdf2-f2d945050386 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.117128] env[61824]: DEBUG nova.compute.provider_tree [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.210552] env[61824]: DEBUG nova.network.neutron [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.356101] env[61824]: DEBUG nova.compute.manager [req-bd9dd0d4-6274-4af7-8afb-ce7f840b0258 req-58a93896-7962-4384-9bda-7eb2cd21e225 service nova] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Received event network-vif-deleted-fc168f10-6129-41b7-a5e2-5b3c50580726 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.594614] env[61824]: INFO nova.compute.manager [-] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Took 1.04 seconds to deallocate network for instance. [ 602.595102] env[61824]: DEBUG nova.compute.claims [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.595278] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.597281] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.619584] env[61824]: DEBUG nova.scheduler.client.report [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.626078] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.626313] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.626466] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.626655] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.626813] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.626959] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.627380] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.627519] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.627944] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.628207] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.628430] env[61824]: DEBUG nova.virt.hardware [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.629350] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4baf02d-e760-4fef-9e34-ad31c0bdf66e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.641194] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d167c39-4a36-43ba-8f20-41f091c496e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.713665] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ac34084-fc72-4c9d-8932-368923e444e8 req-40fd3b65-d226-4f37-b3c0-c5218e04c779 service nova] Releasing lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.854504] env[61824]: DEBUG nova.compute.manager [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Received event network-changed-2e470afc-43b7-4ca4-959e-f759868f109f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.854735] env[61824]: DEBUG nova.compute.manager [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Refreshing instance network info cache due to event network-changed-2e470afc-43b7-4ca4-959e-f759868f109f. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 602.854900] env[61824]: DEBUG oslo_concurrency.lockutils [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] Acquiring lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.855067] env[61824]: DEBUG oslo_concurrency.lockutils [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] Acquired lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.855238] env[61824]: DEBUG nova.network.neutron [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Refreshing network info cache for port 2e470afc-43b7-4ca4-959e-f759868f109f {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 603.134607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.576s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.135135] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 603.140590] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.888s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.182882] env[61824]: ERROR nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. [ 603.182882] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.182882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.182882] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.182882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.182882] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.182882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.182882] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.182882] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.182882] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 603.182882] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.182882] env[61824]: ERROR nova.compute.manager raise self.value [ 603.182882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.182882] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.182882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.182882] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.183338] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.183338] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.183338] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. [ 603.183338] env[61824]: ERROR nova.compute.manager [ 603.183338] env[61824]: Traceback (most recent call last): [ 603.183338] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.183338] env[61824]: listener.cb(fileno) [ 603.183338] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.183338] env[61824]: result = function(*args, **kwargs) [ 603.183338] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.183338] env[61824]: return func(*args, **kwargs) [ 603.183338] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.183338] env[61824]: raise e [ 603.183338] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.183338] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 603.183338] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.183338] env[61824]: created_port_ids = self._update_ports_for_instance( [ 603.183338] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.183338] env[61824]: with excutils.save_and_reraise_exception(): [ 603.183338] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.183338] env[61824]: self.force_reraise() [ 603.183338] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.183338] env[61824]: raise self.value [ 603.183338] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.183338] env[61824]: updated_port = self._update_port( [ 603.183338] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.183338] env[61824]: _ensure_no_port_binding_failure(port) [ 603.183338] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.183338] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.184169] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. [ 603.184169] env[61824]: Removing descriptor: 19 [ 603.184169] env[61824]: ERROR nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Traceback (most recent call last): [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] yield resources [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self.driver.spawn(context, instance, image_meta, [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.184169] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] vm_ref = self.build_virtual_machine(instance, [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] for vif in network_info: [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return self._sync_wrapper(fn, *args, **kwargs) [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self.wait() [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self[:] = self._gt.wait() [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return self._exit_event.wait() [ 603.184478] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] result = hub.switch() [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return self.greenlet.switch() [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] result = function(*args, **kwargs) [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return func(*args, **kwargs) [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] raise e [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] nwinfo = self.network_api.allocate_for_instance( [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.184818] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] created_port_ids = self._update_ports_for_instance( [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] with excutils.save_and_reraise_exception(): [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self.force_reraise() [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] raise self.value [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] updated_port = self._update_port( [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] _ensure_no_port_binding_failure(port) [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.185354] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] raise exception.PortBindingFailed(port_id=port['id']) [ 603.185688] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] nova.exception.PortBindingFailed: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. [ 603.185688] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] [ 603.185688] env[61824]: INFO nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Terminating instance [ 603.186411] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Acquiring lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.376870] env[61824]: DEBUG nova.network.neutron [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.472952] env[61824]: DEBUG nova.network.neutron [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.645465] env[61824]: DEBUG nova.compute.utils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.651939] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.651939] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 603.700460] env[61824]: DEBUG nova.policy [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '316c089785a44aa897040f3c0ef7a071', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0924bfa41b5e42d68818557d2b0ace7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.974199] env[61824]: DEBUG oslo_concurrency.lockutils [req-233f9381-7a7d-4066-ae41-32abe1626358 req-bb7fe893-158c-49fe-8ef4-f5688eeefbc1 service nova] Releasing lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.975687] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Acquired lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.975898] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.119247] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4200faee-17da-47a5-9b83-0b22d023e07e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.129257] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c810560-2be5-4054-bcde-878ae02abd0e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.168665] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 604.172418] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1651a05-662c-4e1a-b238-f0cf2729cf0c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.181396] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39922f40-44b1-46cc-9f74-722d941eae5b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.196960] env[61824]: DEBUG nova.compute.provider_tree [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.199349] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Successfully created port: 5329488e-40e5-4a16-9f55-a0eb1e897477 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.361814] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Acquiring lock "d889d359-f767-4853-8092-070e7656c284" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.362058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Lock "d889d359-f767-4853-8092-070e7656c284" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.501950] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.580098] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.703858] env[61824]: DEBUG nova.scheduler.client.report [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.928130] env[61824]: DEBUG nova.compute.manager [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Received event network-vif-deleted-2e470afc-43b7-4ca4-959e-f759868f109f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.928130] env[61824]: DEBUG nova.compute.manager [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Received event network-changed-5329488e-40e5-4a16-9f55-a0eb1e897477 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.928130] env[61824]: DEBUG nova.compute.manager [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Refreshing instance network info cache due to event network-changed-5329488e-40e5-4a16-9f55-a0eb1e897477. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 604.928130] env[61824]: DEBUG oslo_concurrency.lockutils [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] Acquiring lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.928766] env[61824]: DEBUG oslo_concurrency.lockutils [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] Acquired lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.929171] env[61824]: DEBUG nova.network.neutron [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Refreshing network info cache for port 5329488e-40e5-4a16-9f55-a0eb1e897477 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 605.054905] env[61824]: ERROR nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. [ 605.054905] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.054905] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.054905] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.054905] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.054905] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.054905] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.054905] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.054905] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.054905] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 605.054905] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.054905] env[61824]: ERROR nova.compute.manager raise self.value [ 605.054905] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.054905] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.054905] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.054905] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.055571] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.055571] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.055571] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. [ 605.055571] env[61824]: ERROR nova.compute.manager [ 605.055571] env[61824]: Traceback (most recent call last): [ 605.055571] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.055571] env[61824]: listener.cb(fileno) [ 605.055571] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.055571] env[61824]: result = function(*args, **kwargs) [ 605.055571] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.055571] env[61824]: return func(*args, **kwargs) [ 605.055571] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.055571] env[61824]: raise e [ 605.055571] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.055571] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 605.055571] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.055571] env[61824]: created_port_ids = self._update_ports_for_instance( [ 605.055571] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.055571] env[61824]: with excutils.save_and_reraise_exception(): [ 605.055571] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.055571] env[61824]: self.force_reraise() [ 605.055571] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.055571] env[61824]: raise self.value [ 605.055571] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.055571] env[61824]: updated_port = self._update_port( [ 605.055571] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.055571] env[61824]: _ensure_no_port_binding_failure(port) [ 605.055571] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.055571] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.056229] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. [ 605.056229] env[61824]: Removing descriptor: 19 [ 605.082654] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Releasing lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.083351] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 605.083351] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 605.083710] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bdea5a48-5441-4a5f-8c56-03b52b18f6dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.092753] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a08acf-9f55-42b3-a534-a0b1b3a78775 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.113678] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0 could not be found. [ 605.113932] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 605.114124] env[61824]: INFO nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 605.114367] env[61824]: DEBUG oslo.service.loopingcall [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 605.114582] env[61824]: DEBUG nova.compute.manager [-] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.114676] env[61824]: DEBUG nova.network.neutron [-] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.129949] env[61824]: DEBUG nova.network.neutron [-] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.183708] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 605.208249] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 605.208497] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 605.208649] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 605.208825] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 605.208965] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 605.209122] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 605.209317] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 605.209470] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 605.209630] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 605.209789] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 605.209956] env[61824]: DEBUG nova.virt.hardware [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 605.210675] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.073s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.211279] env[61824]: ERROR nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Traceback (most recent call last): [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self.driver.spawn(context, instance, image_meta, [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] vm_ref = self.build_virtual_machine(instance, [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.211279] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] for vif in network_info: [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return self._sync_wrapper(fn, *args, **kwargs) [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self.wait() [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self[:] = self._gt.wait() [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return self._exit_event.wait() [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] result = hub.switch() [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.211544] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return self.greenlet.switch() [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] result = function(*args, **kwargs) [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] return func(*args, **kwargs) [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] raise e [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] nwinfo = self.network_api.allocate_for_instance( [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] created_port_ids = self._update_ports_for_instance( [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] with excutils.save_and_reraise_exception(): [ 605.211852] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] self.force_reraise() [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] raise self.value [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] updated_port = self._update_port( [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] _ensure_no_port_binding_failure(port) [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] raise exception.PortBindingFailed(port_id=port['id']) [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] nova.exception.PortBindingFailed: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. [ 605.212153] env[61824]: ERROR nova.compute.manager [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] [ 605.212472] env[61824]: DEBUG nova.compute.utils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.213752] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc6a029-2c8c-47c4-aa61-b763b853f4e3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.216934] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.563s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.222692] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Build of instance 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e was re-scheduled: Binding failed for port ef6f7ba7-75a8-4bdd-89bc-7c56cd4b5137, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 605.222692] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 605.222692] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Acquiring lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.222692] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Acquired lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.222918] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.227559] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d449b8-ac4a-42af-ac2b-5a77d0b2ca00 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.241771] env[61824]: ERROR nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Traceback (most recent call last): [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] yield resources [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self.driver.spawn(context, instance, image_meta, [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] vm_ref = self.build_virtual_machine(instance, [ 605.241771] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] for vif in network_info: [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] return self._sync_wrapper(fn, *args, **kwargs) [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self.wait() [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self[:] = self._gt.wait() [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] return self._exit_event.wait() [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 605.242138] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] current.throw(*self._exc) [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] result = function(*args, **kwargs) [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] return func(*args, **kwargs) [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] raise e [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] nwinfo = self.network_api.allocate_for_instance( [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] created_port_ids = self._update_ports_for_instance( [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] with excutils.save_and_reraise_exception(): [ 605.242443] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self.force_reraise() [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] raise self.value [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] updated_port = self._update_port( [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] _ensure_no_port_binding_failure(port) [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] raise exception.PortBindingFailed(port_id=port['id']) [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] nova.exception.PortBindingFailed: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. [ 605.242721] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] [ 605.242721] env[61824]: INFO nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Terminating instance [ 605.243850] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.448912] env[61824]: DEBUG nova.network.neutron [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.520601] env[61824]: DEBUG nova.network.neutron [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.632395] env[61824]: DEBUG nova.network.neutron [-] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.742349] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.817205] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.022792] env[61824]: DEBUG oslo_concurrency.lockutils [req-03db93aa-badc-48e9-9480-1e48f1dc0591 req-a30db834-51f9-4e87-bf33-88b83c9bb5f5 service nova] Releasing lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.023206] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.023372] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 606.081498] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7323d45-8434-4a36-9c53-be9f2680cd13 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.088850] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181476d3-2b44-4291-8243-66487f19bc70 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.118499] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8d39fd-8ec1-4396-a31b-cd7f137354f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.125461] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc954fb-de5c-458f-b391-467677b6bc8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.138614] env[61824]: INFO nova.compute.manager [-] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Took 1.02 seconds to deallocate network for instance. [ 606.139108] env[61824]: DEBUG nova.compute.provider_tree [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.144902] env[61824]: DEBUG nova.compute.claims [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 606.145106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.320287] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Releasing lock "refresh_cache-9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.320544] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 606.320728] env[61824]: DEBUG nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.320898] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.335249] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.542218] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.613452] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.642372] env[61824]: DEBUG nova.scheduler.client.report [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.837977] env[61824]: DEBUG nova.network.neutron [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.967822] env[61824]: DEBUG nova.compute.manager [req-6ad52226-17dc-4c12-96af-7f957a7b97a1 req-98a88eff-cfa9-4715-9079-150f280fd3f5 service nova] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Received event network-vif-deleted-5329488e-40e5-4a16-9f55-a0eb1e897477 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 607.116384] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.116895] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 607.117097] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.117398] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9f485900-c153-4f67-9eea-6c5d3e488332 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.126272] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd002017-2e46-438f-aa23-bfdb068b022f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.147357] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.147956] env[61824]: ERROR nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Traceback (most recent call last): [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self.driver.spawn(context, instance, image_meta, [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] vm_ref = self.build_virtual_machine(instance, [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.147956] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] for vif in network_info: [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return self._sync_wrapper(fn, *args, **kwargs) [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self.wait() [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self[:] = self._gt.wait() [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return self._exit_event.wait() [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] result = hub.switch() [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.148281] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return self.greenlet.switch() [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] result = function(*args, **kwargs) [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] return func(*args, **kwargs) [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] raise e [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] nwinfo = self.network_api.allocate_for_instance( [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] created_port_ids = self._update_ports_for_instance( [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] with excutils.save_and_reraise_exception(): [ 607.148638] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] self.force_reraise() [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] raise self.value [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] updated_port = self._update_port( [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] _ensure_no_port_binding_failure(port) [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] raise exception.PortBindingFailed(port_id=port['id']) [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] nova.exception.PortBindingFailed: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. [ 607.148992] env[61824]: ERROR nova.compute.manager [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] [ 607.149299] env[61824]: DEBUG nova.compute.utils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.149973] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 35c3e1eb-6b98-42f7-8f92-420560feba92 could not be found. [ 607.150378] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.150378] env[61824]: INFO nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Took 0.03 seconds to destroy the instance on the hypervisor. [ 607.150561] env[61824]: DEBUG oslo.service.loopingcall [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.150980] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Build of instance ed4352e2-3321-4a05-8a3f-a02e375ecbfb was re-scheduled: Binding failed for port eb95b184-7f01-4fbd-a154-7cc225b378c1, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 607.151379] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 607.151592] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Acquiring lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.151733] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Acquired lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.151934] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.152892] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.142s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.155406] env[61824]: DEBUG nova.compute.manager [-] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 607.155406] env[61824]: DEBUG nova.network.neutron [-] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 607.170094] env[61824]: DEBUG nova.network.neutron [-] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.340500] env[61824]: INFO nova.compute.manager [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] [instance: 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e] Took 1.02 seconds to deallocate network for instance. [ 607.672381] env[61824]: DEBUG nova.network.neutron [-] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.676193] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.723313] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.044621] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fa888b-12d2-4a1a-bfb9-edbebc1ff300 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.052186] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77881e4-2679-4816-9d73-3387e35fff30 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.083039] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f57d39c-8934-419f-ba1e-67c986fa7693 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.090278] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85cdb356-10d9-4d83-bf1f-e96624b3e5fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.102928] env[61824]: DEBUG nova.compute.provider_tree [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.174646] env[61824]: INFO nova.compute.manager [-] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Took 1.02 seconds to deallocate network for instance. [ 608.176979] env[61824]: DEBUG nova.compute.claims [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.177209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.226019] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Releasing lock "refresh_cache-ed4352e2-3321-4a05-8a3f-a02e375ecbfb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.226275] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 608.226438] env[61824]: DEBUG nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.226604] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.243040] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.368248] env[61824]: INFO nova.scheduler.client.report [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Deleted allocations for instance 9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e [ 608.605803] env[61824]: DEBUG nova.scheduler.client.report [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.746222] env[61824]: DEBUG nova.network.neutron [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.875462] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3b2eaa9-89ff-45bf-a977-448dd9976a67 tempest-ServerTagsTestJSON-1308813039 tempest-ServerTagsTestJSON-1308813039-project-member] Lock "9b87db2b-e7b0-4677-b70f-f2fb1d6d3c4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.142s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.110539] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.111299] env[61824]: ERROR nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Traceback (most recent call last): [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self.driver.spawn(context, instance, image_meta, [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] vm_ref = self.build_virtual_machine(instance, [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.111299] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] for vif in network_info: [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return self._sync_wrapper(fn, *args, **kwargs) [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self.wait() [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self[:] = self._gt.wait() [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return self._exit_event.wait() [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] result = hub.switch() [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.111583] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return self.greenlet.switch() [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] result = function(*args, **kwargs) [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] return func(*args, **kwargs) [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] raise e [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] nwinfo = self.network_api.allocate_for_instance( [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] created_port_ids = self._update_ports_for_instance( [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] with excutils.save_and_reraise_exception(): [ 609.111914] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] self.force_reraise() [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] raise self.value [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] updated_port = self._update_port( [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] _ensure_no_port_binding_failure(port) [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] raise exception.PortBindingFailed(port_id=port['id']) [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] nova.exception.PortBindingFailed: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. [ 609.112271] env[61824]: ERROR nova.compute.manager [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] [ 609.112563] env[61824]: DEBUG nova.compute.utils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.113740] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Build of instance 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c was re-scheduled: Binding failed for port 4c302980-ed9e-4f69-a400-e95a9be397a2, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.114171] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.114426] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.115028] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.115028] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.115764] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.007s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.117647] env[61824]: INFO nova.compute.claims [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.248604] env[61824]: INFO nova.compute.manager [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] [instance: ed4352e2-3321-4a05-8a3f-a02e375ecbfb] Took 1.02 seconds to deallocate network for instance. [ 609.377933] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 609.656700] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.813125] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.904024] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.281569] env[61824]: INFO nova.scheduler.client.report [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Deleted allocations for instance ed4352e2-3321-4a05-8a3f-a02e375ecbfb [ 610.315874] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "refresh_cache-5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.316157] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.316353] env[61824]: DEBUG nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.316534] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.339124] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.688907] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8d7525-b455-432a-9753-b8f9432f20d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.701778] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1998181c-5b89-483c-be11-4b626290ef5a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.740722] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cc1658-94dd-4aab-84f4-7d7af56e88ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.749652] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aff1745-170c-46a5-a5d5-1d57ad3b7162 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.768040] env[61824]: DEBUG nova.compute.provider_tree [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.792435] env[61824]: DEBUG oslo_concurrency.lockutils [None req-45dfc316-f4a3-4b4f-9331-cb24e44096dd tempest-ServerExternalEventsTest-1813911304 tempest-ServerExternalEventsTest-1813911304-project-member] Lock "ed4352e2-3321-4a05-8a3f-a02e375ecbfb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.309s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.845395] env[61824]: DEBUG nova.network.neutron [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.271045] env[61824]: DEBUG nova.scheduler.client.report [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.295483] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 611.348337] env[61824]: INFO nova.compute.manager [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c] Took 1.03 seconds to deallocate network for instance. [ 611.776173] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.660s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.776173] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.778859] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.221s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.780624] env[61824]: INFO nova.compute.claims [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.822450] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.286023] env[61824]: DEBUG nova.compute.utils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 612.288845] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 612.289025] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 612.338094] env[61824]: DEBUG nova.policy [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '19cf885c5fd14cefb56c34719f21412d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '875df0be2fe34b5f96d86a312c2dd4f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 612.378899] env[61824]: INFO nova.scheduler.client.report [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted allocations for instance 5c80ad4d-0fa1-4003-a6a1-2c9039ad679c [ 612.640032] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Successfully created port: 905f14c2-ec78-472f-a1ad-a2fc7cf297e7 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.790299] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.887423] env[61824]: DEBUG oslo_concurrency.lockutils [None req-53ed46a1-97d6-4419-9ef3-447249ea8e0d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5c80ad4d-0fa1-4003-a6a1-2c9039ad679c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.234s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.203283] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d590f8-ac57-4d0d-bae9-9d4d947feaf7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.211225] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534292c8-e5b9-4a21-9e77-ced5c2e69383 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.242898] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46ed09b-8e8a-4df7-9d0a-cf02e5c0a590 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.250418] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c5e71a-caa2-49dd-8a28-1212b08544ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.263926] env[61824]: DEBUG nova.compute.provider_tree [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.302756] env[61824]: INFO nova.virt.block_device [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Booting with volume 6f74117f-fafe-42fa-a813-b60e123de966 at /dev/sda [ 613.355918] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb5e1830-2c00-4ab7-b5a2-d6676a617afb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.368402] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322846eb-eed6-497a-803b-2f0d9a8dcd95 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.393191] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.396740] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-333d1fcd-1879-49a7-b7ad-d87473a84da6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.406423] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07419ee8-608d-4f7a-954d-eeb91436804a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.054979] env[61824]: DEBUG nova.scheduler.client.report [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.064882] env[61824]: ERROR nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. [ 614.064882] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.064882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.064882] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.064882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.064882] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.064882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.064882] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.064882] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.064882] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 614.064882] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.064882] env[61824]: ERROR nova.compute.manager raise self.value [ 614.064882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.064882] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.064882] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.064882] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.065290] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.065290] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.065290] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. [ 614.065290] env[61824]: ERROR nova.compute.manager [ 614.065290] env[61824]: Traceback (most recent call last): [ 614.065290] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.065290] env[61824]: listener.cb(fileno) [ 614.065290] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.065290] env[61824]: result = function(*args, **kwargs) [ 614.065290] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.065290] env[61824]: return func(*args, **kwargs) [ 614.065290] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.065290] env[61824]: raise e [ 614.065290] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.065290] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 614.065290] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.065290] env[61824]: created_port_ids = self._update_ports_for_instance( [ 614.065290] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.065290] env[61824]: with excutils.save_and_reraise_exception(): [ 614.065290] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.065290] env[61824]: self.force_reraise() [ 614.065290] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.065290] env[61824]: raise self.value [ 614.065290] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.065290] env[61824]: updated_port = self._update_port( [ 614.065290] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.065290] env[61824]: _ensure_no_port_binding_failure(port) [ 614.065290] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.065290] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.065974] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. [ 614.065974] env[61824]: Removing descriptor: 19 [ 614.066431] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91befecb-c520-4e90-bc2b-cb7fbe0feb59 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.069673] env[61824]: DEBUG nova.compute.manager [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Received event network-changed-905f14c2-ec78-472f-a1ad-a2fc7cf297e7 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 614.069849] env[61824]: DEBUG nova.compute.manager [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Refreshing instance network info cache due to event network-changed-905f14c2-ec78-472f-a1ad-a2fc7cf297e7. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 614.070066] env[61824]: DEBUG oslo_concurrency.lockutils [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] Acquiring lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.070205] env[61824]: DEBUG oslo_concurrency.lockutils [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] Acquired lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.070355] env[61824]: DEBUG nova.network.neutron [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Refreshing network info cache for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 614.074334] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "61d1f216-7ff1-419d-b415-3dce3537b1b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.074334] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "61d1f216-7ff1-419d-b415-3dce3537b1b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.079694] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affb868c-54a0-456d-8250-e6799114e825 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.087293] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.093650] env[61824]: DEBUG nova.virt.block_device [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Updating existing volume attachment record: 82e451ce-cd17-44cf-b3b0-674cbcc92360 {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 614.561436] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.782s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.561963] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.564723] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.754s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.609607] env[61824]: DEBUG nova.network.neutron [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.745016] env[61824]: DEBUG nova.network.neutron [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.077331] env[61824]: DEBUG nova.compute.utils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.077331] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.077331] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 615.122528] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.122528] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.160433] env[61824]: DEBUG nova.policy [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fab4d7e1cb4da9a54c55cc9762c329', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d19b5b4169b4a55b1d14216da62a659', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.245291] env[61824]: DEBUG oslo_concurrency.lockutils [req-495b95da-6250-454d-8b6d-1f037c7d8841 req-40580532-7755-470e-a1ae-bcc8b6cda2ff service nova] Releasing lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.529359] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4ec7e8-3382-4fc3-aa69-eb3a95101add {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.539489] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-488ae5b3-acaa-4477-9a24-ab276df11e2c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.585534] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.585534] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7effe21-c1e9-4872-ac76-7ca2ef3520fe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.595522] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7206f969-d72e-4c4e-843b-9dde6eec6ad3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.605015] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Successfully created port: f55003cb-b5e8-41d9-b965-67cd8403fa9e {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.619678] env[61824]: DEBUG nova.compute.provider_tree [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.622652] env[61824]: DEBUG nova.compute.manager [req-f70f7cec-a985-4bc7-b704-4a9f324b9ba6 req-803a4197-537e-4ccd-b444-45c3a3213187 service nova] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Received event network-vif-deleted-905f14c2-ec78-472f-a1ad-a2fc7cf297e7 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.633547] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 615.633547] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 615.633547] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Rebuilding the list of instances to heal {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 616.124515] env[61824]: DEBUG nova.scheduler.client.report [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.134079] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.134242] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.135939] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.135939] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.135939] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.135939] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 616.165039] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "refresh_cache-de5a46be-193a-40d8-aeba-d5d4de24ef95" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.165213] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquired lock "refresh_cache-de5a46be-193a-40d8-aeba-d5d4de24ef95" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.165362] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Forcefully refreshing network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 616.165516] env[61824]: DEBUG nova.objects.instance [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lazy-loading 'info_cache' on Instance uuid de5a46be-193a-40d8-aeba-d5d4de24ef95 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 616.229852] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.230021] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.230241] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.230395] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.230572] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.230786] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.230850] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.231932] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.231932] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.231932] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.231932] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.231932] env[61824]: DEBUG nova.virt.hardware [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.232905] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549b7071-9d65-4af5-87ce-24ddb6563a94 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.245052] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ceefc87-6846-4c7b-bf29-7f47a443388b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.258147] env[61824]: ERROR nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Traceback (most recent call last): [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] yield resources [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self.driver.spawn(context, instance, image_meta, [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] vm_ref = self.build_virtual_machine(instance, [ 616.258147] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] for vif in network_info: [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] return self._sync_wrapper(fn, *args, **kwargs) [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self.wait() [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self[:] = self._gt.wait() [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] return self._exit_event.wait() [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 616.258515] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] current.throw(*self._exc) [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] result = function(*args, **kwargs) [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] return func(*args, **kwargs) [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] raise e [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] nwinfo = self.network_api.allocate_for_instance( [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] created_port_ids = self._update_ports_for_instance( [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] with excutils.save_and_reraise_exception(): [ 616.258906] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self.force_reraise() [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] raise self.value [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] updated_port = self._update_port( [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] _ensure_no_port_binding_failure(port) [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] raise exception.PortBindingFailed(port_id=port['id']) [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] nova.exception.PortBindingFailed: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. [ 616.259427] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] [ 616.259427] env[61824]: INFO nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Terminating instance [ 616.259997] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Acquiring lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.260177] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Acquired lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.260339] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.573430] env[61824]: ERROR nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. [ 616.573430] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 616.573430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.573430] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 616.573430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.573430] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 616.573430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.573430] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 616.573430] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.573430] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 616.573430] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.573430] env[61824]: ERROR nova.compute.manager raise self.value [ 616.573430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.573430] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 616.573430] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.573430] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 616.574030] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.574030] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 616.574030] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. [ 616.574030] env[61824]: ERROR nova.compute.manager [ 616.574030] env[61824]: Traceback (most recent call last): [ 616.574030] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 616.574030] env[61824]: listener.cb(fileno) [ 616.574030] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.574030] env[61824]: result = function(*args, **kwargs) [ 616.574030] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.574030] env[61824]: return func(*args, **kwargs) [ 616.574030] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.574030] env[61824]: raise e [ 616.574030] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.574030] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 616.574030] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.574030] env[61824]: created_port_ids = self._update_ports_for_instance( [ 616.574030] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.574030] env[61824]: with excutils.save_and_reraise_exception(): [ 616.574030] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.574030] env[61824]: self.force_reraise() [ 616.574030] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.574030] env[61824]: raise self.value [ 616.574030] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.574030] env[61824]: updated_port = self._update_port( [ 616.574030] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.574030] env[61824]: _ensure_no_port_binding_failure(port) [ 616.574030] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.574030] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 616.574687] env[61824]: nova.exception.PortBindingFailed: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. [ 616.574687] env[61824]: Removing descriptor: 19 [ 616.593047] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.627998] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.628135] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.628297] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.628475] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.628808] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.628808] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.628949] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.629115] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.629282] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.629441] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.629608] env[61824]: DEBUG nova.virt.hardware [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.630388] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.066s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.630985] env[61824]: ERROR nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Traceback (most recent call last): [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self.driver.spawn(context, instance, image_meta, [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] vm_ref = self.build_virtual_machine(instance, [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.630985] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] for vif in network_info: [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return self._sync_wrapper(fn, *args, **kwargs) [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self.wait() [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self[:] = self._gt.wait() [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return self._exit_event.wait() [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] result = hub.switch() [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.631351] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return self.greenlet.switch() [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] result = function(*args, **kwargs) [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] return func(*args, **kwargs) [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] raise e [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] nwinfo = self.network_api.allocate_for_instance( [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] created_port_ids = self._update_ports_for_instance( [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] with excutils.save_and_reraise_exception(): [ 616.632156] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] self.force_reraise() [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] raise self.value [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] updated_port = self._update_port( [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] _ensure_no_port_binding_failure(port) [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] raise exception.PortBindingFailed(port_id=port['id']) [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] nova.exception.PortBindingFailed: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. [ 616.632821] env[61824]: ERROR nova.compute.manager [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] [ 616.633196] env[61824]: DEBUG nova.compute.utils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.633299] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797a8411-80d3-4789-8c11-27b99c71d00b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.636194] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Build of instance 507a2e12-55a1-4d2c-b3d2-577914cc245e was re-scheduled: Binding failed for port 9f2d1282-fd12-432d-9101-81ae66dcec8e, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 616.636626] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 616.636835] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Acquiring lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.636978] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Acquired lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.637199] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.638122] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.106s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.646683] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ce2b1e-9bc5-418d-8984-4f5d32a2cc16 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.662024] env[61824]: ERROR nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Traceback (most recent call last): [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] yield resources [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self.driver.spawn(context, instance, image_meta, [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] vm_ref = self.build_virtual_machine(instance, [ 616.662024] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] for vif in network_info: [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] return self._sync_wrapper(fn, *args, **kwargs) [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self.wait() [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self[:] = self._gt.wait() [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] return self._exit_event.wait() [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 616.662474] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] current.throw(*self._exc) [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] result = function(*args, **kwargs) [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] return func(*args, **kwargs) [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] raise e [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] nwinfo = self.network_api.allocate_for_instance( [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] created_port_ids = self._update_ports_for_instance( [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] with excutils.save_and_reraise_exception(): [ 616.662774] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self.force_reraise() [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] raise self.value [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] updated_port = self._update_port( [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] _ensure_no_port_binding_failure(port) [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] raise exception.PortBindingFailed(port_id=port['id']) [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] nova.exception.PortBindingFailed: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. [ 616.663107] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] [ 616.663107] env[61824]: INFO nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Terminating instance [ 616.664498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.664586] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquired lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.664763] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.779199] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.860635] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.165499] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.208176] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.216280] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.257381] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.294129] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.363094] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Releasing lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.363631] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 617.363937] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a1dc8ba-640a-4d97-b5cc-c7498f1d6dfc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.374582] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1225b5-ec92-4040-be33-fea134637dcc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.396722] env[61824]: WARNING nova.virt.vmwareapi.driver [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 002a7370-024d-4959-a56d-1a13c260f60e could not be found. [ 617.396970] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 617.399835] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dbe551c4-56fe-462a-afe2-22330ca50081 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.407678] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9bad377-197b-4987-8f54-36beff49ffad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.432545] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 002a7370-024d-4959-a56d-1a13c260f60e could not be found. [ 617.432965] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 617.432965] env[61824]: INFO nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Took 0.07 seconds to destroy the instance on the hypervisor. [ 617.433207] env[61824]: DEBUG oslo.service.loopingcall [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.435862] env[61824]: DEBUG nova.compute.manager [-] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.435983] env[61824]: DEBUG nova.network.neutron [-] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.469208] env[61824]: DEBUG nova.network.neutron [-] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.573689] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7734caf5-7a28-4610-9e7f-dfc6e76439b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.580943] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02575d28-efac-4b83-a375-a4ca1c58e46b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.610773] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fbd3301-84ca-4520-9b42-a5f38477e28e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.618927] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c53e7bd-a428-47ec-a2e5-247a243d8e31 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.632871] env[61824]: DEBUG nova.compute.provider_tree [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.728635] env[61824]: DEBUG nova.compute.manager [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Received event network-changed-f55003cb-b5e8-41d9-b965-67cd8403fa9e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.728823] env[61824]: DEBUG nova.compute.manager [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Refreshing instance network info cache due to event network-changed-f55003cb-b5e8-41d9-b965-67cd8403fa9e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 617.729033] env[61824]: DEBUG oslo_concurrency.lockutils [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] Acquiring lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.757909] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Releasing lock "refresh_cache-507a2e12-55a1-4d2c-b3d2-577914cc245e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.757909] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 617.757909] env[61824]: DEBUG nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.757909] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.773031] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.797185] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Releasing lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.797642] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 617.797877] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 617.798761] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.799789] env[61824]: DEBUG oslo_concurrency.lockutils [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] Acquired lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.799967] env[61824]: DEBUG nova.network.neutron [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Refreshing network info cache for port f55003cb-b5e8-41d9-b965-67cd8403fa9e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 617.800837] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5eeb4e16-d122-45a3-982f-2f87f718c3f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.811662] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5ab8ce-28b4-4c60-83a3-a8df3102dcf9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.835325] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 46b484d0-1e87-4438-9791-e8199c76d21f could not be found. [ 617.835551] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 617.835734] env[61824]: INFO nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 617.835940] env[61824]: DEBUG oslo.service.loopingcall [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.836142] env[61824]: DEBUG nova.compute.manager [-] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 617.836228] env[61824]: DEBUG nova.network.neutron [-] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.854267] env[61824]: DEBUG nova.network.neutron [-] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.971895] env[61824]: DEBUG nova.network.neutron [-] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.136512] env[61824]: DEBUG nova.scheduler.client.report [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.277029] env[61824]: DEBUG nova.network.neutron [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.304360] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Releasing lock "refresh_cache-de5a46be-193a-40d8-aeba-d5d4de24ef95" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.304578] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Updated the network info_cache for instance {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 618.306728] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.307176] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.307642] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.307809] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.307982] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.308157] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.308444] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 618.308444] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.334519] env[61824]: DEBUG nova.network.neutron [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.356651] env[61824]: DEBUG nova.network.neutron [-] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.401950] env[61824]: DEBUG nova.network.neutron [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.474886] env[61824]: INFO nova.compute.manager [-] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Took 1.04 seconds to deallocate network for instance. [ 618.643151] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.643151] env[61824]: ERROR nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. [ 618.643151] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Traceback (most recent call last): [ 618.643151] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.643151] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self.driver.spawn(context, instance, image_meta, [ 618.643151] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.643151] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.643151] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.643151] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] vm_ref = self.build_virtual_machine(instance, [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] for vif in network_info: [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return self._sync_wrapper(fn, *args, **kwargs) [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self.wait() [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self[:] = self._gt.wait() [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return self._exit_event.wait() [ 618.643566] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] result = hub.switch() [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return self.greenlet.switch() [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] result = function(*args, **kwargs) [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] return func(*args, **kwargs) [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] raise e [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] nwinfo = self.network_api.allocate_for_instance( [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.643856] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] created_port_ids = self._update_ports_for_instance( [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] with excutils.save_and_reraise_exception(): [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] self.force_reraise() [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] raise self.value [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] updated_port = self._update_port( [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] _ensure_no_port_binding_failure(port) [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.644169] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] raise exception.PortBindingFailed(port_id=port['id']) [ 618.644447] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] nova.exception.PortBindingFailed: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. [ 618.644447] env[61824]: ERROR nova.compute.manager [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] [ 618.644447] env[61824]: DEBUG nova.compute.utils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.644530] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.452s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.644740] env[61824]: DEBUG nova.objects.instance [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lazy-loading 'resources' on Instance uuid de5a46be-193a-40d8-aeba-d5d4de24ef95 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 618.645988] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Build of instance 51d54995-0d13-4560-bdd6-7d9f9f5210dd was re-scheduled: Binding failed for port 1ebad03f-4822-4162-a19e-499ae24a52e7, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.646423] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.646641] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Acquiring lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.646784] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Acquired lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.646964] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.779895] env[61824]: INFO nova.compute.manager [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] [instance: 507a2e12-55a1-4d2c-b3d2-577914cc245e] Took 1.02 seconds to deallocate network for instance. [ 618.811629] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.861505] env[61824]: INFO nova.compute.manager [-] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Took 1.02 seconds to deallocate network for instance. [ 618.861682] env[61824]: DEBUG nova.compute.claims [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 618.861860] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.905210] env[61824]: DEBUG oslo_concurrency.lockutils [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] Releasing lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.905584] env[61824]: DEBUG nova.compute.manager [req-91076c2b-4849-4a27-9241-44df4a13aa1f req-309f156f-5c4b-46f7-bc85-7aa42f725520 service nova] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Received event network-vif-deleted-f55003cb-b5e8-41d9-b965-67cd8403fa9e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.038474] env[61824]: INFO nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Took 0.56 seconds to detach 1 volumes for instance. [ 619.043914] env[61824]: DEBUG nova.compute.claims [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.044097] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.330886] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.514797] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.526985] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb33ed34-52c3-4709-b299-adb425bb603f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.536191] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0208bf7f-afb8-4ca2-9aae-630d6d944a57 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.570805] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39af12cf-55c8-41ca-be00-e059c02395b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.580270] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0d54bb-4984-477c-bf5c-f2af36ce36eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.594695] env[61824]: DEBUG nova.compute.provider_tree [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.823172] env[61824]: INFO nova.scheduler.client.report [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Deleted allocations for instance 507a2e12-55a1-4d2c-b3d2-577914cc245e [ 620.021338] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Releasing lock "refresh_cache-51d54995-0d13-4560-bdd6-7d9f9f5210dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.021600] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.021784] env[61824]: DEBUG nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.021955] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 620.046459] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.098856] env[61824]: DEBUG nova.scheduler.client.report [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.332791] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d7595074-462f-4d96-ba1d-c5b1894626f9 tempest-ServerActionsTestOtherB-1564700025 tempest-ServerActionsTestOtherB-1564700025-project-member] Lock "507a2e12-55a1-4d2c-b3d2-577914cc245e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.337s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.549586] env[61824]: DEBUG nova.network.neutron [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.604618] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.607951] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.012s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.647073] env[61824]: INFO nova.scheduler.client.report [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Deleted allocations for instance de5a46be-193a-40d8-aeba-d5d4de24ef95 [ 620.835066] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 621.052300] env[61824]: INFO nova.compute.manager [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] [instance: 51d54995-0d13-4560-bdd6-7d9f9f5210dd] Took 1.03 seconds to deallocate network for instance. [ 621.163702] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9d9df0da-6caa-43cf-adc2-db697905dc25 tempest-ServersAaction247Test-404114046 tempest-ServersAaction247Test-404114046-project-member] Lock "de5a46be-193a-40d8-aeba-d5d4de24ef95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.699s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.354023] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.507168] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62ca60c-aea4-498e-abac-eedcd476f6c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.514348] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d560bf-bfd0-4b6a-815a-6df038bc4ee3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.544306] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab1d9e3-e3ba-4cf1-8957-7cac7db9f769 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.551892] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b358610-c3ec-4d74-8b19-a40601561419 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.570122] env[61824]: DEBUG nova.compute.provider_tree [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.079661] env[61824]: DEBUG nova.scheduler.client.report [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.107231] env[61824]: INFO nova.scheduler.client.report [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Deleted allocations for instance 51d54995-0d13-4560-bdd6-7d9f9f5210dd [ 622.587028] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.587693] env[61824]: ERROR nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Traceback (most recent call last): [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self.driver.spawn(context, instance, image_meta, [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] vm_ref = self.build_virtual_machine(instance, [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.587693] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] for vif in network_info: [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return self._sync_wrapper(fn, *args, **kwargs) [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self.wait() [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self[:] = self._gt.wait() [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return self._exit_event.wait() [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] result = hub.switch() [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.588044] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return self.greenlet.switch() [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] result = function(*args, **kwargs) [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] return func(*args, **kwargs) [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] raise e [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] nwinfo = self.network_api.allocate_for_instance( [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] created_port_ids = self._update_ports_for_instance( [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] with excutils.save_and_reraise_exception(): [ 622.588399] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] self.force_reraise() [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] raise self.value [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] updated_port = self._update_port( [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] _ensure_no_port_binding_failure(port) [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] raise exception.PortBindingFailed(port_id=port['id']) [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] nova.exception.PortBindingFailed: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. [ 622.588751] env[61824]: ERROR nova.compute.manager [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] [ 622.589083] env[61824]: DEBUG nova.compute.utils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.589758] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.445s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.596267] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Build of instance e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb was re-scheduled: Binding failed for port fc168f10-6129-41b7-a5e2-5b3c50580726, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.596968] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.596968] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Acquiring lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.596968] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Acquired lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.596968] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.617107] env[61824]: DEBUG oslo_concurrency.lockutils [None req-38b7ae34-11ac-4b3d-a70f-ba15bb81614e tempest-ServersTestManualDisk-1999569825 tempest-ServersTestManualDisk-1999569825-project-member] Lock "51d54995-0d13-4560-bdd6-7d9f9f5210dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.034s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.118171] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 623.127100] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.258603] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.518374] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb3a8ee-616a-40f8-ab78-cc804f2646c5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.530063] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d2ec72-ac46-4bf4-919e-b8dc46efe54e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.564996] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9185162d-1525-42ba-9bd1-4b85dd46caa3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.572445] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab50518-5489-4c68-be43-cf65534c22ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.587009] env[61824]: DEBUG nova.compute.provider_tree [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.641258] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.762853] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Releasing lock "refresh_cache-e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.763984] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.763984] env[61824]: DEBUG nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.763984] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.782654] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.092299] env[61824]: DEBUG nova.scheduler.client.report [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.285626] env[61824]: DEBUG nova.network.neutron [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.604015] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.606499] env[61824]: ERROR nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Traceback (most recent call last): [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self.driver.spawn(context, instance, image_meta, [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] vm_ref = self.build_virtual_machine(instance, [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.606499] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] for vif in network_info: [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return self._sync_wrapper(fn, *args, **kwargs) [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self.wait() [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self[:] = self._gt.wait() [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return self._exit_event.wait() [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] result = hub.switch() [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.606980] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return self.greenlet.switch() [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] result = function(*args, **kwargs) [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] return func(*args, **kwargs) [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] raise e [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] nwinfo = self.network_api.allocate_for_instance( [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] created_port_ids = self._update_ports_for_instance( [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] with excutils.save_and_reraise_exception(): [ 624.607285] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] self.force_reraise() [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] raise self.value [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] updated_port = self._update_port( [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] _ensure_no_port_binding_failure(port) [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] raise exception.PortBindingFailed(port_id=port['id']) [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] nova.exception.PortBindingFailed: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. [ 624.607579] env[61824]: ERROR nova.compute.manager [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] [ 624.607829] env[61824]: DEBUG nova.compute.utils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.607829] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.430s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.610752] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Build of instance 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0 was re-scheduled: Binding failed for port 2e470afc-43b7-4ca4-959e-f759868f109f, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.611347] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.611347] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Acquiring lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.611734] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Acquired lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.612747] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 624.790268] env[61824]: INFO nova.compute.manager [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] [instance: e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb] Took 1.03 seconds to deallocate network for instance. [ 625.147791] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.281314] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.517725] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197cfcbe-dc97-4206-8d1e-c2f456d60619 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.528235] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d92c73-5127-4dce-b0db-088a30bbd5b6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.556967] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dab4b43-c8f2-4a7b-acd0-6470b6618d7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.564713] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184ded73-85cc-4d26-89b2-1b98c749f388 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.578113] env[61824]: DEBUG nova.compute.provider_tree [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.787294] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Releasing lock "refresh_cache-21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.787548] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.787734] env[61824]: DEBUG nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.788103] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.816484] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.857130] env[61824]: INFO nova.scheduler.client.report [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Deleted allocations for instance e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb [ 626.081014] env[61824]: DEBUG nova.scheduler.client.report [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.320664] env[61824]: DEBUG nova.network.neutron [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.366810] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd2bf802-c486-45fa-b047-6ba19f74c41c tempest-ServersTestJSON-911656621 tempest-ServersTestJSON-911656621-project-member] Lock "e7ba3a8d-4bd0-4d78-861f-866ce2ade0cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.266s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.590594] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.591646] env[61824]: ERROR nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Traceback (most recent call last): [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self.driver.spawn(context, instance, image_meta, [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] vm_ref = self.build_virtual_machine(instance, [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.591646] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] for vif in network_info: [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] return self._sync_wrapper(fn, *args, **kwargs) [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self.wait() [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self[:] = self._gt.wait() [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] return self._exit_event.wait() [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] current.throw(*self._exc) [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.594348] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] result = function(*args, **kwargs) [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] return func(*args, **kwargs) [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] raise e [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] nwinfo = self.network_api.allocate_for_instance( [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] created_port_ids = self._update_ports_for_instance( [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] with excutils.save_and_reraise_exception(): [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] self.force_reraise() [ 626.594938] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] raise self.value [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] updated_port = self._update_port( [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] _ensure_no_port_binding_failure(port) [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] raise exception.PortBindingFailed(port_id=port['id']) [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] nova.exception.PortBindingFailed: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. [ 626.595282] env[61824]: ERROR nova.compute.manager [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] [ 626.595282] env[61824]: DEBUG nova.compute.utils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 626.595709] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.690s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.598756] env[61824]: INFO nova.compute.claims [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.601023] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Build of instance 35c3e1eb-6b98-42f7-8f92-420560feba92 was re-scheduled: Binding failed for port 5329488e-40e5-4a16-9f55-a0eb1e897477, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 626.601023] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 626.601023] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.601023] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.601434] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 626.822547] env[61824]: INFO nova.compute.manager [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] [instance: 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0] Took 1.03 seconds to deallocate network for instance. [ 626.873184] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.129727] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.306447] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.396523] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.809911] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-35c3e1eb-6b98-42f7-8f92-420560feba92" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.810221] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 627.810656] env[61824]: DEBUG nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 627.810656] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 627.868435] env[61824]: INFO nova.scheduler.client.report [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Deleted allocations for instance 21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0 [ 627.971152] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.022220] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af660e8-08dc-45d2-b6e9-4d26bfc3efac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.030095] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a12e9e7-38d7-4df7-ae29-232b73f45cd6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.061405] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d38343-a7a9-47df-9ba0-bdd03ce9e6f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.070691] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afa7332-cdda-4ce8-9acb-71a47b239578 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.083205] env[61824]: DEBUG nova.compute.provider_tree [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.379187] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8b69261c-2485-45e0-848e-5f691d21b8af tempest-ServerRescueTestJSON-59388324 tempest-ServerRescueTestJSON-59388324-project-member] Lock "21ad1f7b-cc59-4ad4-a0c0-9e867b20e9f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.369s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.471587] env[61824]: DEBUG nova.network.neutron [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.589687] env[61824]: DEBUG nova.scheduler.client.report [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.883217] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.974608] env[61824]: INFO nova.compute.manager [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 35c3e1eb-6b98-42f7-8f92-420560feba92] Took 1.16 seconds to deallocate network for instance. [ 629.065422] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Acquiring lock "329b3d0d-7007-40e3-a554-fbd0c7497b69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.065560] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Lock "329b3d0d-7007-40e3-a554-fbd0c7497b69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.095233] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.095233] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.097099] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.275s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.102820] env[61824]: INFO nova.compute.claims [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.414387] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.609112] env[61824]: DEBUG nova.compute.utils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.614057] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.614299] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 629.753021] env[61824]: DEBUG nova.policy [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2dd621c2963444e1a73a9c22c92bd481', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bac9aecd7fc4b6aaa87004c017dd2bb', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.011982] env[61824]: INFO nova.scheduler.client.report [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleted allocations for instance 35c3e1eb-6b98-42f7-8f92-420560feba92 [ 630.127275] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.507294] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Successfully created port: 138652a9-ac69-46e0-ab61-047a785a8e9b {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.520429] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ded33fa-859b-486b-9de5-537853b3bff3 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "35c3e1eb-6b98-42f7-8f92-420560feba92" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.703s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.545225] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b53025-d237-4c5a-b28a-93a95b3cd179 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.553352] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0766e7db-84b8-466f-b379-8c07c900ed60 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.587148] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2286908-05b8-4784-b306-f8e911f8f8da {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.597180] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8505cf-3ec4-46ed-b64b-0dd3d794dc44 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.610140] env[61824]: DEBUG nova.compute.provider_tree [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.023604] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.112832] env[61824]: DEBUG nova.scheduler.client.report [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.136948] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.167095] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.167095] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.167095] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.167263] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.167263] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.167263] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.167784] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.168078] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.168393] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.168701] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.169030] env[61824]: DEBUG nova.virt.hardware [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.170455] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57f77d2-3a89-418d-9b75-69c0e0458a31 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.183640] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39411ca4-07f3-4e79-9ee0-df7a5127d0bf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.560045] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.622216] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.622741] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 631.627562] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.540s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.629116] env[61824]: INFO nova.compute.claims [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.750523] env[61824]: DEBUG nova.compute.manager [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Received event network-changed-138652a9-ac69-46e0-ab61-047a785a8e9b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.750729] env[61824]: DEBUG nova.compute.manager [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Refreshing instance network info cache due to event network-changed-138652a9-ac69-46e0-ab61-047a785a8e9b. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.751017] env[61824]: DEBUG oslo_concurrency.lockutils [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] Acquiring lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.751209] env[61824]: DEBUG oslo_concurrency.lockutils [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] Acquired lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.751433] env[61824]: DEBUG nova.network.neutron [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Refreshing network info cache for port 138652a9-ac69-46e0-ab61-047a785a8e9b {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 632.032411] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.032411] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.129362] env[61824]: DEBUG nova.compute.utils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.130809] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 632.130809] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 632.186325] env[61824]: ERROR nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. [ 632.186325] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.186325] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.186325] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.186325] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.186325] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.186325] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.186325] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.186325] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.186325] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 632.186325] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.186325] env[61824]: ERROR nova.compute.manager raise self.value [ 632.186325] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.186325] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.186325] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.186325] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.186895] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.186895] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.186895] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. [ 632.186895] env[61824]: ERROR nova.compute.manager [ 632.186895] env[61824]: Traceback (most recent call last): [ 632.186895] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.186895] env[61824]: listener.cb(fileno) [ 632.186895] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.186895] env[61824]: result = function(*args, **kwargs) [ 632.186895] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.186895] env[61824]: return func(*args, **kwargs) [ 632.186895] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.186895] env[61824]: raise e [ 632.186895] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.186895] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 632.186895] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.186895] env[61824]: created_port_ids = self._update_ports_for_instance( [ 632.186895] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.186895] env[61824]: with excutils.save_and_reraise_exception(): [ 632.186895] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.186895] env[61824]: self.force_reraise() [ 632.186895] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.186895] env[61824]: raise self.value [ 632.186895] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.186895] env[61824]: updated_port = self._update_port( [ 632.186895] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.186895] env[61824]: _ensure_no_port_binding_failure(port) [ 632.186895] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.186895] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.187779] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. [ 632.187779] env[61824]: Removing descriptor: 17 [ 632.187779] env[61824]: ERROR nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Traceback (most recent call last): [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] yield resources [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self.driver.spawn(context, instance, image_meta, [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.187779] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] vm_ref = self.build_virtual_machine(instance, [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] for vif in network_info: [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return self._sync_wrapper(fn, *args, **kwargs) [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self.wait() [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self[:] = self._gt.wait() [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return self._exit_event.wait() [ 632.188134] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] result = hub.switch() [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return self.greenlet.switch() [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] result = function(*args, **kwargs) [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return func(*args, **kwargs) [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] raise e [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] nwinfo = self.network_api.allocate_for_instance( [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.188558] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] created_port_ids = self._update_ports_for_instance( [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] with excutils.save_and_reraise_exception(): [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self.force_reraise() [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] raise self.value [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] updated_port = self._update_port( [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] _ensure_no_port_binding_failure(port) [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.188919] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] raise exception.PortBindingFailed(port_id=port['id']) [ 632.189289] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] nova.exception.PortBindingFailed: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. [ 632.189289] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] [ 632.189289] env[61824]: INFO nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Terminating instance [ 632.190668] env[61824]: DEBUG nova.policy [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e0382e0563054c1086db47868ae9e10a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5ac68c40b5dd42ef9639b511a981bbd2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 632.192531] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Acquiring lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.274194] env[61824]: DEBUG nova.network.neutron [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.346478] env[61824]: DEBUG nova.network.neutron [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.417616] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "9fd8e236-3556-4b95-952f-3c324b896e29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.417907] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "9fd8e236-3556-4b95-952f-3c324b896e29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.633744] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.731229] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Successfully created port: 09195547-1b85-432a-8167-473f4d76bd77 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.849282] env[61824]: DEBUG oslo_concurrency.lockutils [req-4454e43c-856a-4d4c-a83e-6c141a1b8278 req-91e52c76-6f5f-4d38-93e6-8a10749cdbd9 service nova] Releasing lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.849690] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Acquired lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.849873] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 633.028386] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3cff42-c3d3-4a6a-9930-d785f097bd19 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.036881] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a225841-f174-4302-a0ea-f005453fcf87 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.072267] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207396d4-9c8c-4b45-801a-ce285f1ca67f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.080352] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d0f341-1d8e-43d0-8c20-d4c4d9999272 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.094370] env[61824]: DEBUG nova.compute.provider_tree [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.386750] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.514644] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.601230] env[61824]: DEBUG nova.scheduler.client.report [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.642518] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.675414] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.675725] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.675903] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.676132] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.676287] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.676434] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.676634] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.676789] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.676951] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.677124] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.677293] env[61824]: DEBUG nova.virt.hardware [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.678142] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6122b9d4-1592-44a2-af6b-5b681e0a0328 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.687226] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f622d676-23cf-47cb-a46a-84cbd41b7a36 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.958119] env[61824]: ERROR nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. [ 633.958119] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.958119] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.958119] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.958119] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.958119] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.958119] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.958119] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.958119] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.958119] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 633.958119] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.958119] env[61824]: ERROR nova.compute.manager raise self.value [ 633.958119] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.958119] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.958119] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.958119] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.958574] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.958574] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.958574] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. [ 633.958574] env[61824]: ERROR nova.compute.manager [ 633.958574] env[61824]: Traceback (most recent call last): [ 633.958574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.958574] env[61824]: listener.cb(fileno) [ 633.958574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.958574] env[61824]: result = function(*args, **kwargs) [ 633.958574] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.958574] env[61824]: return func(*args, **kwargs) [ 633.958574] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.958574] env[61824]: raise e [ 633.958574] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.958574] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 633.958574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.958574] env[61824]: created_port_ids = self._update_ports_for_instance( [ 633.958574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.958574] env[61824]: with excutils.save_and_reraise_exception(): [ 633.958574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.958574] env[61824]: self.force_reraise() [ 633.958574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.958574] env[61824]: raise self.value [ 633.958574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.958574] env[61824]: updated_port = self._update_port( [ 633.958574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.958574] env[61824]: _ensure_no_port_binding_failure(port) [ 633.958574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.958574] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.959243] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. [ 633.959243] env[61824]: Removing descriptor: 19 [ 633.959243] env[61824]: ERROR nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] Traceback (most recent call last): [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] yield resources [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self.driver.spawn(context, instance, image_meta, [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.959243] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] vm_ref = self.build_virtual_machine(instance, [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] for vif in network_info: [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return self._sync_wrapper(fn, *args, **kwargs) [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self.wait() [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self[:] = self._gt.wait() [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return self._exit_event.wait() [ 633.959587] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] result = hub.switch() [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return self.greenlet.switch() [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] result = function(*args, **kwargs) [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return func(*args, **kwargs) [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] raise e [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] nwinfo = self.network_api.allocate_for_instance( [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.959890] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] created_port_ids = self._update_ports_for_instance( [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] with excutils.save_and_reraise_exception(): [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self.force_reraise() [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] raise self.value [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] updated_port = self._update_port( [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] _ensure_no_port_binding_failure(port) [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.960215] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] raise exception.PortBindingFailed(port_id=port['id']) [ 633.960491] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] nova.exception.PortBindingFailed: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. [ 633.960491] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] [ 633.960491] env[61824]: INFO nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Terminating instance [ 633.963893] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Acquiring lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.963893] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Acquired lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.963893] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.017208] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Releasing lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.017951] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.017951] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.018505] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30cb0a64-834e-4ffd-a48b-d383632aaea8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.028692] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0f3d62-c839-4305-a48d-638bf6499f47 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.050925] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95635ec3-cbfd-4dc0-a595-02062a0bda6e could not be found. [ 634.051186] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.051364] env[61824]: INFO nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 634.051681] env[61824]: DEBUG oslo.service.loopingcall [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.051889] env[61824]: DEBUG nova.compute.manager [-] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.051986] env[61824]: DEBUG nova.network.neutron [-] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 634.067784] env[61824]: DEBUG nova.network.neutron [-] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.108021] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.108021] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.109150] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 15.298s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.109481] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.113316] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 634.113688] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.252s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.120807] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e9e76d8-cfbd-4ca0-a492-bb9623f08b23 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.124968] env[61824]: DEBUG nova.compute.manager [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Received event network-changed-09195547-1b85-432a-8167-473f4d76bd77 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.125329] env[61824]: DEBUG nova.compute.manager [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Refreshing instance network info cache due to event network-changed-09195547-1b85-432a-8167-473f4d76bd77. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 634.125704] env[61824]: DEBUG oslo_concurrency.lockutils [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] Acquiring lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.132146] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e41160a-e7d4-43ea-9cd9-4e822d75a88b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.148023] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbac4783-24b6-4224-84bd-c5ae72dd0fc5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.156920] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cb72f6-c71b-4823-b68c-fb0c13d5d488 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.190957] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181513MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 634.191143] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.193318] env[61824]: DEBUG nova.compute.manager [req-e358b3c2-8f02-4af5-9253-283e75b65dbd req-dafe592a-696e-46dc-874c-0cc922601724 service nova] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Received event network-vif-deleted-138652a9-ac69-46e0-ab61-047a785a8e9b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.478722] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "34f72773-5950-4710-b3e8-5e5c4877f227" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.478957] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "34f72773-5950-4710-b3e8-5e5c4877f227" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.480650] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.532963] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.569213] env[61824]: DEBUG nova.network.neutron [-] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.621102] env[61824]: DEBUG nova.compute.utils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.625340] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.625508] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 634.667623] env[61824]: DEBUG nova.policy [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0e58836210684edc8323383720196f0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9be0d9c3b42498485929655a2ea1d04', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.847924] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Acquiring lock "0f86f95d-2c41-47b2-93f8-c7e6983c913a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.848206] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Lock "0f86f95d-2c41-47b2-93f8-c7e6983c913a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.961843] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Successfully created port: 5910859c-229c-4d45-b859-61b6bafce7e8 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.986083] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a0bbbf-dfc0-460e-aca2-74377cc4f32e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.993870] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168b8a79-ffe3-4645-95b6-1ba7b02e197d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.025359] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eb56f7-04de-4fd6-b44a-98ed82a11610 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.032656] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a020013c-dd98-412c-9329-040331e5a6f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.036583] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Releasing lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.036997] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 635.037216] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 635.037491] env[61824]: DEBUG oslo_concurrency.lockutils [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] Acquired lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.037650] env[61824]: DEBUG nova.network.neutron [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Refreshing network info cache for port 09195547-1b85-432a-8167-473f4d76bd77 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 635.038680] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-407c1d7b-5619-467b-8729-51ac1de126b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.051654] env[61824]: DEBUG nova.compute.provider_tree [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.057099] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-029bc6ed-290f-4969-96ee-43edfe03039d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.071228] env[61824]: INFO nova.compute.manager [-] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Took 1.02 seconds to deallocate network for instance. [ 635.073380] env[61824]: DEBUG nova.compute.claims [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.073558] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.079641] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 496f00ef-7487-45d9-9356-529f9390544f could not be found. [ 635.079850] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 635.080038] env[61824]: INFO nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 635.080290] env[61824]: DEBUG oslo.service.loopingcall [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.080485] env[61824]: DEBUG nova.compute.manager [-] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.080577] env[61824]: DEBUG nova.network.neutron [-] [instance: 496f00ef-7487-45d9-9356-529f9390544f] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.102741] env[61824]: DEBUG nova.network.neutron [-] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.128897] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.554869] env[61824]: DEBUG nova.network.neutron [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.557507] env[61824]: DEBUG nova.scheduler.client.report [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.604801] env[61824]: DEBUG nova.network.neutron [-] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.645436] env[61824]: DEBUG nova.network.neutron [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.800048] env[61824]: ERROR nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. [ 635.800048] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.800048] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.800048] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.800048] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.800048] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.800048] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.800048] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.800048] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.800048] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 635.800048] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.800048] env[61824]: ERROR nova.compute.manager raise self.value [ 635.800048] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.800048] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.800048] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.800048] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.800570] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.800570] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.800570] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. [ 635.800570] env[61824]: ERROR nova.compute.manager [ 635.800570] env[61824]: Traceback (most recent call last): [ 635.800570] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.800570] env[61824]: listener.cb(fileno) [ 635.800570] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.800570] env[61824]: result = function(*args, **kwargs) [ 635.800570] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.800570] env[61824]: return func(*args, **kwargs) [ 635.800570] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.800570] env[61824]: raise e [ 635.800570] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.800570] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 635.800570] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.800570] env[61824]: created_port_ids = self._update_ports_for_instance( [ 635.800570] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.800570] env[61824]: with excutils.save_and_reraise_exception(): [ 635.800570] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.800570] env[61824]: self.force_reraise() [ 635.800570] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.800570] env[61824]: raise self.value [ 635.800570] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.800570] env[61824]: updated_port = self._update_port( [ 635.800570] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.800570] env[61824]: _ensure_no_port_binding_failure(port) [ 635.800570] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.800570] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.801373] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. [ 635.801373] env[61824]: Removing descriptor: 19 [ 636.067042] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.067042] env[61824]: ERROR nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. [ 636.067042] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Traceback (most recent call last): [ 636.067042] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.067042] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self.driver.spawn(context, instance, image_meta, [ 636.067042] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 636.067042] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.067042] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.067042] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] vm_ref = self.build_virtual_machine(instance, [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] for vif in network_info: [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] return self._sync_wrapper(fn, *args, **kwargs) [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self.wait() [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self[:] = self._gt.wait() [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] return self._exit_event.wait() [ 636.067571] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] current.throw(*self._exc) [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] result = function(*args, **kwargs) [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] return func(*args, **kwargs) [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] raise e [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] nwinfo = self.network_api.allocate_for_instance( [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] created_port_ids = self._update_ports_for_instance( [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.067967] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] with excutils.save_and_reraise_exception(): [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] self.force_reraise() [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] raise self.value [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] updated_port = self._update_port( [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] _ensure_no_port_binding_failure(port) [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] raise exception.PortBindingFailed(port_id=port['id']) [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] nova.exception.PortBindingFailed: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. [ 636.068662] env[61824]: ERROR nova.compute.manager [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] [ 636.069072] env[61824]: DEBUG nova.compute.utils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.069072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.025s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.071590] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Build of instance 46b484d0-1e87-4438-9791-e8199c76d21f was re-scheduled: Binding failed for port f55003cb-b5e8-41d9-b965-67cd8403fa9e, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.072032] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.072222] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquiring lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.072369] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Acquired lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.072524] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 636.107115] env[61824]: INFO nova.compute.manager [-] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Took 1.03 seconds to deallocate network for instance. [ 636.109224] env[61824]: DEBUG nova.compute.claims [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 636.109399] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.139484] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.144572] env[61824]: DEBUG nova.compute.manager [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Received event network-vif-deleted-09195547-1b85-432a-8167-473f4d76bd77 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.144746] env[61824]: DEBUG nova.compute.manager [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Received event network-changed-5910859c-229c-4d45-b859-61b6bafce7e8 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 636.144876] env[61824]: DEBUG nova.compute.manager [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Refreshing instance network info cache due to event network-changed-5910859c-229c-4d45-b859-61b6bafce7e8. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 636.145093] env[61824]: DEBUG oslo_concurrency.lockutils [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] Acquiring lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.145237] env[61824]: DEBUG oslo_concurrency.lockutils [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] Acquired lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.145394] env[61824]: DEBUG nova.network.neutron [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Refreshing network info cache for port 5910859c-229c-4d45-b859-61b6bafce7e8 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 636.147403] env[61824]: DEBUG oslo_concurrency.lockutils [req-a58b9451-271e-4341-98b0-d902fd2e05f3 req-16ff1073-5e03-4baa-9245-20ba33250429 service nova] Releasing lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.164214] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.164491] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.164692] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.164908] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.165118] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.165297] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.165530] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.165719] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.165916] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.166190] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.166405] env[61824]: DEBUG nova.virt.hardware [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.167527] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee30232-b9fa-47df-97e8-466f97c23ccc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.177040] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f565e6a-6294-4448-96f3-547065acc651 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.191154] env[61824]: ERROR nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Traceback (most recent call last): [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] yield resources [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self.driver.spawn(context, instance, image_meta, [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] vm_ref = self.build_virtual_machine(instance, [ 636.191154] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] for vif in network_info: [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] return self._sync_wrapper(fn, *args, **kwargs) [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self.wait() [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self[:] = self._gt.wait() [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] return self._exit_event.wait() [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 636.191487] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] current.throw(*self._exc) [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] result = function(*args, **kwargs) [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] return func(*args, **kwargs) [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] raise e [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] nwinfo = self.network_api.allocate_for_instance( [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] created_port_ids = self._update_ports_for_instance( [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] with excutils.save_and_reraise_exception(): [ 636.191788] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self.force_reraise() [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] raise self.value [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] updated_port = self._update_port( [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] _ensure_no_port_binding_failure(port) [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] raise exception.PortBindingFailed(port_id=port['id']) [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] nova.exception.PortBindingFailed: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. [ 636.192364] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] [ 636.192364] env[61824]: INFO nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Terminating instance [ 636.193287] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Acquiring lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.593114] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.642476] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.665091] env[61824]: DEBUG nova.network.neutron [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.744121] env[61824]: DEBUG nova.network.neutron [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.930967] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50a103d-fa15-4bb4-981a-186d3a42328e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.939036] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6132e2f8-0a57-474b-b5fc-4046435f5451 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.969083] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a581bd59-1ff9-4e52-b7bd-972e19e3caf4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.976859] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55b2871-d2e1-4628-aebe-4a279abb80f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.989159] env[61824]: DEBUG nova.compute.provider_tree [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.145090] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Releasing lock "refresh_cache-46b484d0-1e87-4438-9791-e8199c76d21f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.145347] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 637.145511] env[61824]: DEBUG nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.145675] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 637.160955] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.248292] env[61824]: DEBUG oslo_concurrency.lockutils [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] Releasing lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.248615] env[61824]: DEBUG nova.compute.manager [req-e372c80c-7bbe-4fef-81ce-1dfc9a67e7ef req-b77f8d30-7592-4e6d-b991-0e7797d4218b service nova] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Received event network-vif-deleted-5910859c-229c-4d45-b859-61b6bafce7e8 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.249030] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Acquired lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.249247] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 637.492039] env[61824]: DEBUG nova.scheduler.client.report [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.663550] env[61824]: DEBUG nova.network.neutron [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.766057] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.836699] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.997057] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.997612] env[61824]: ERROR nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Traceback (most recent call last): [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self.driver.spawn(context, instance, image_meta, [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] vm_ref = self.build_virtual_machine(instance, [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.997612] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] for vif in network_info: [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] return self._sync_wrapper(fn, *args, **kwargs) [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self.wait() [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self[:] = self._gt.wait() [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] return self._exit_event.wait() [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] current.throw(*self._exc) [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.998030] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] result = function(*args, **kwargs) [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] return func(*args, **kwargs) [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] raise e [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] nwinfo = self.network_api.allocate_for_instance( [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] created_port_ids = self._update_ports_for_instance( [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] with excutils.save_and_reraise_exception(): [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] self.force_reraise() [ 637.998382] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] raise self.value [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] updated_port = self._update_port( [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] _ensure_no_port_binding_failure(port) [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] raise exception.PortBindingFailed(port_id=port['id']) [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] nova.exception.PortBindingFailed: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. [ 637.998753] env[61824]: ERROR nova.compute.manager [instance: 002a7370-024d-4959-a56d-1a13c260f60e] [ 637.998753] env[61824]: DEBUG nova.compute.utils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.999733] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.646s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.001402] env[61824]: INFO nova.compute.claims [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.003860] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Build of instance 002a7370-024d-4959-a56d-1a13c260f60e was re-scheduled: Binding failed for port 905f14c2-ec78-472f-a1ad-a2fc7cf297e7, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.004333] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.004984] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Acquiring lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.004984] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Acquired lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.004984] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 638.166717] env[61824]: INFO nova.compute.manager [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] [instance: 46b484d0-1e87-4438-9791-e8199c76d21f] Took 1.02 seconds to deallocate network for instance. [ 638.339397] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Releasing lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.339830] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.340041] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.340348] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e65cf31-c239-4d19-9b02-61291b3c4954 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.349368] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddae2c3-2aaf-468b-84b3-f84d54b36c50 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.369926] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad02e0c7-3d12-48ad-a393-26749fd8c502 could not be found. [ 638.370149] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 638.370331] env[61824]: INFO nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Took 0.03 seconds to destroy the instance on the hypervisor. [ 638.370561] env[61824]: DEBUG oslo.service.loopingcall [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.370766] env[61824]: DEBUG nova.compute.manager [-] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.370854] env[61824]: DEBUG nova.network.neutron [-] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.385185] env[61824]: DEBUG nova.network.neutron [-] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.524497] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.596139] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.887649] env[61824]: DEBUG nova.network.neutron [-] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.099349] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Releasing lock "refresh_cache-002a7370-024d-4959-a56d-1a13c260f60e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.099599] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.099798] env[61824]: DEBUG nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.099968] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 639.119202] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.196194] env[61824]: INFO nova.scheduler.client.report [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Deleted allocations for instance 46b484d0-1e87-4438-9791-e8199c76d21f [ 639.379230] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35561f05-c9d2-4ea7-b109-c04f45ec0c4f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.387135] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db1f34b-922d-491e-8cec-da5c3a6909a3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.390422] env[61824]: INFO nova.compute.manager [-] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Took 1.02 seconds to deallocate network for instance. [ 639.392584] env[61824]: DEBUG nova.compute.claims [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.392756] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.417767] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ac32a1-abd6-4038-88e0-5e7159072253 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.424824] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cf98e93-012a-4f50-887c-98d3a9ca2f8d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.437651] env[61824]: DEBUG nova.compute.provider_tree [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.622025] env[61824]: DEBUG nova.network.neutron [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.706714] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ee90a45-8c73-45ca-aea9-dd8af87f5860 tempest-DeleteServersAdminTestJSON-165745837 tempest-DeleteServersAdminTestJSON-165745837-project-member] Lock "46b484d0-1e87-4438-9791-e8199c76d21f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.019s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.941353] env[61824]: DEBUG nova.scheduler.client.report [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.124703] env[61824]: INFO nova.compute.manager [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] [instance: 002a7370-024d-4959-a56d-1a13c260f60e] Took 1.02 seconds to deallocate network for instance. [ 640.209623] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.446682] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.447224] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.449744] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.809s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.451133] env[61824]: INFO nova.compute.claims [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.735840] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.955930] env[61824]: DEBUG nova.compute.utils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.960339] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.960518] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 641.000164] env[61824]: DEBUG nova.policy [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a849332813d74fdb9d374eb0547efa0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1393b3c0605c4f139bf37f80d85bbe4d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 641.152040] env[61824]: INFO nova.scheduler.client.report [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Deleted allocations for instance 002a7370-024d-4959-a56d-1a13c260f60e [ 641.255481] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Successfully created port: 1a7fc60a-1f66-40ad-b353-e5d0a2d77403 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.461018] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.661353] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efe8af77-3727-43be-a8a1-c5454c19284d tempest-ServerActionsV293TestJSON-1721592387 tempest-ServerActionsV293TestJSON-1721592387-project-member] Lock "002a7370-024d-4959-a56d-1a13c260f60e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.743s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.856701] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2263fd2a-60be-439c-97e1-54b77b6a89b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.865141] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082cab94-96cb-4967-964b-ff6e77398e6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.905372] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8957e90e-dd0d-44d5-8a25-1e1d633123f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.913402] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82533a67-d731-456e-9ff9-a2cfa575dcf1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.929300] env[61824]: DEBUG nova.compute.provider_tree [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.163057] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.435193] env[61824]: DEBUG nova.scheduler.client.report [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.470416] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.503640] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.503888] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.504075] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.504275] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.504417] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.504560] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.504764] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.504919] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.505095] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.505261] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.505431] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.506315] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f956d7c-8ba0-4fab-8f11-999cbcd18ca8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.516189] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ef9868-5a6d-4a98-b8e1-d943064f5004 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.533694] env[61824]: DEBUG nova.compute.manager [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Received event network-changed-1a7fc60a-1f66-40ad-b353-e5d0a2d77403 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 642.533889] env[61824]: DEBUG nova.compute.manager [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Refreshing instance network info cache due to event network-changed-1a7fc60a-1f66-40ad-b353-e5d0a2d77403. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 642.534149] env[61824]: DEBUG oslo_concurrency.lockutils [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] Acquiring lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.534299] env[61824]: DEBUG oslo_concurrency.lockutils [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] Acquired lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.536029] env[61824]: DEBUG nova.network.neutron [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Refreshing network info cache for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 642.685593] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.786056] env[61824]: ERROR nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. [ 642.786056] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.786056] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.786056] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.786056] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.786056] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.786056] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.786056] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.786056] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.786056] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 642.786056] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.786056] env[61824]: ERROR nova.compute.manager raise self.value [ 642.786056] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.786056] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.786056] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.786056] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.786507] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.786507] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.786507] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. [ 642.786507] env[61824]: ERROR nova.compute.manager [ 642.786507] env[61824]: Traceback (most recent call last): [ 642.786507] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.786507] env[61824]: listener.cb(fileno) [ 642.786507] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.786507] env[61824]: result = function(*args, **kwargs) [ 642.786507] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.786507] env[61824]: return func(*args, **kwargs) [ 642.786507] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.786507] env[61824]: raise e [ 642.786507] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.786507] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 642.786507] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.786507] env[61824]: created_port_ids = self._update_ports_for_instance( [ 642.786507] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.786507] env[61824]: with excutils.save_and_reraise_exception(): [ 642.786507] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.786507] env[61824]: self.force_reraise() [ 642.786507] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.786507] env[61824]: raise self.value [ 642.786507] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.786507] env[61824]: updated_port = self._update_port( [ 642.786507] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.786507] env[61824]: _ensure_no_port_binding_failure(port) [ 642.786507] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.786507] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.787308] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. [ 642.787308] env[61824]: Removing descriptor: 19 [ 642.787308] env[61824]: ERROR nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Traceback (most recent call last): [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] yield resources [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self.driver.spawn(context, instance, image_meta, [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.787308] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] vm_ref = self.build_virtual_machine(instance, [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] for vif in network_info: [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return self._sync_wrapper(fn, *args, **kwargs) [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self.wait() [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self[:] = self._gt.wait() [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return self._exit_event.wait() [ 642.787654] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] result = hub.switch() [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return self.greenlet.switch() [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] result = function(*args, **kwargs) [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return func(*args, **kwargs) [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] raise e [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] nwinfo = self.network_api.allocate_for_instance( [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.788092] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] created_port_ids = self._update_ports_for_instance( [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] with excutils.save_and_reraise_exception(): [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self.force_reraise() [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] raise self.value [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] updated_port = self._update_port( [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] _ensure_no_port_binding_failure(port) [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.788449] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] raise exception.PortBindingFailed(port_id=port['id']) [ 642.788761] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] nova.exception.PortBindingFailed: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. [ 642.788761] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] [ 642.788761] env[61824]: INFO nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Terminating instance [ 642.789925] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.944681] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.945277] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.949488] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.553s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.952040] env[61824]: INFO nova.compute.claims [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.061026] env[61824]: DEBUG nova.network.neutron [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.178264] env[61824]: DEBUG nova.network.neutron [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.459873] env[61824]: DEBUG nova.compute.utils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.465379] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.465379] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 643.517648] env[61824]: DEBUG nova.policy [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a849332813d74fdb9d374eb0547efa0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1393b3c0605c4f139bf37f80d85bbe4d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.681913] env[61824]: DEBUG oslo_concurrency.lockutils [req-83626259-56f3-4ff7-a2f4-2db8b9731a2b req-1f6731f1-0d7e-4014-bac7-59728c6e4dec service nova] Releasing lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.682375] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.682559] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.910876] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Successfully created port: 1fd63140-442e-45dc-bd5e-e5ae125899e4 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.969020] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 644.208089] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.369709] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.572560] env[61824]: DEBUG nova.compute.manager [req-5e337c56-21b4-4190-8ed7-c85197885113 req-7d8b9b82-c0fe-47ed-9bd6-b37d3940988a service nova] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Received event network-vif-deleted-1a7fc60a-1f66-40ad-b353-e5d0a2d77403 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.654648] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e2b77b-0b25-4273-a8a4-cb91b2f89274 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.666135] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e142036d-ed52-4a6a-a2a7-4c4914f14666 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.701991] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5e680a-dec5-4e6a-b72c-04f89cc51a51 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.710158] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e535a4a-9c6d-48bd-8315-6c22772dc4d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.725846] env[61824]: DEBUG nova.compute.provider_tree [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.872081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.872553] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 644.872744] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.873055] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77ce80b0-dfee-47d7-bf40-19e005b67b78 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.886144] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfd8943-ddbe-44de-a826-7c24e6d0e9ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.910592] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d51370a2-86d5-4c9e-978c-f3318acfd3a7 could not be found. [ 644.910892] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.911118] env[61824]: INFO nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 644.911409] env[61824]: DEBUG oslo.service.loopingcall [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.911665] env[61824]: DEBUG nova.compute.manager [-] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.911787] env[61824]: DEBUG nova.network.neutron [-] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.979880] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 645.008291] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.008555] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.008761] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.008961] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.009162] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.009349] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.009592] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.009908] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.010622] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.010719] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.010934] env[61824]: DEBUG nova.virt.hardware [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.011964] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35bd587c-e1d0-4b21-83e8-c02b3b7eae96 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.020536] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d12ed9a-75df-46c7-942b-ed7d3139a7c2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.065255] env[61824]: DEBUG nova.network.neutron [-] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.077124] env[61824]: ERROR nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. [ 645.077124] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.077124] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.077124] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.077124] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.077124] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.077124] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.077124] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.077124] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.077124] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 645.077124] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.077124] env[61824]: ERROR nova.compute.manager raise self.value [ 645.077124] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.077124] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.077124] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.077124] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.077532] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.077532] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.077532] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. [ 645.077532] env[61824]: ERROR nova.compute.manager [ 645.077532] env[61824]: Traceback (most recent call last): [ 645.077532] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.077532] env[61824]: listener.cb(fileno) [ 645.077532] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.077532] env[61824]: result = function(*args, **kwargs) [ 645.077532] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.077532] env[61824]: return func(*args, **kwargs) [ 645.077532] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.077532] env[61824]: raise e [ 645.077532] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.077532] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 645.077532] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.077532] env[61824]: created_port_ids = self._update_ports_for_instance( [ 645.077532] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.077532] env[61824]: with excutils.save_and_reraise_exception(): [ 645.077532] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.077532] env[61824]: self.force_reraise() [ 645.077532] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.077532] env[61824]: raise self.value [ 645.077532] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.077532] env[61824]: updated_port = self._update_port( [ 645.077532] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.077532] env[61824]: _ensure_no_port_binding_failure(port) [ 645.077532] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.077532] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.078279] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. [ 645.078279] env[61824]: Removing descriptor: 19 [ 645.078279] env[61824]: ERROR nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Traceback (most recent call last): [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] yield resources [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self.driver.spawn(context, instance, image_meta, [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.078279] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] vm_ref = self.build_virtual_machine(instance, [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] for vif in network_info: [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return self._sync_wrapper(fn, *args, **kwargs) [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self.wait() [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self[:] = self._gt.wait() [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return self._exit_event.wait() [ 645.078596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] result = hub.switch() [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return self.greenlet.switch() [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] result = function(*args, **kwargs) [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return func(*args, **kwargs) [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] raise e [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] nwinfo = self.network_api.allocate_for_instance( [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.078941] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] created_port_ids = self._update_ports_for_instance( [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] with excutils.save_and_reraise_exception(): [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self.force_reraise() [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] raise self.value [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] updated_port = self._update_port( [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] _ensure_no_port_binding_failure(port) [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.079266] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] raise exception.PortBindingFailed(port_id=port['id']) [ 645.079570] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] nova.exception.PortBindingFailed: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. [ 645.079570] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] [ 645.079570] env[61824]: INFO nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Terminating instance [ 645.079570] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.079570] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.079570] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 645.229439] env[61824]: DEBUG nova.scheduler.client.report [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.571746] env[61824]: DEBUG nova.network.neutron [-] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.603410] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.678756] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.734973] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.786s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.735935] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.738620] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.325s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.740375] env[61824]: INFO nova.compute.claims [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.075761] env[61824]: INFO nova.compute.manager [-] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Took 1.16 seconds to deallocate network for instance. [ 646.078268] env[61824]: DEBUG nova.compute.claims [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 646.078524] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.184748] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.185219] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 646.185419] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 646.185731] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9161c61c-5c57-4680-b45e-4af2357f1fcf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.196402] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ee94ab-22e2-4b42-868d-2c61c2fe17aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.223057] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 325a641d-2509-49e6-bfbc-0b2cb85c1ff9 could not be found. [ 646.223388] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 646.223647] env[61824]: INFO nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 646.224154] env[61824]: DEBUG oslo.service.loopingcall [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.224447] env[61824]: DEBUG nova.compute.manager [-] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.224590] env[61824]: DEBUG nova.network.neutron [-] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 646.243580] env[61824]: DEBUG nova.network.neutron [-] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.245825] env[61824]: DEBUG nova.compute.utils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 646.250595] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 646.250833] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 646.310797] env[61824]: DEBUG nova.policy [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ab8e3edd4aa24dfd86509610c9d76ff5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e32a3f0d9dba49c990b7d77d9af9a9f9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.615353] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Successfully created port: fec9e102-4433-4cbd-8d36-a1d5dfb250c6 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.632413] env[61824]: DEBUG nova.compute.manager [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Received event network-changed-1fd63140-442e-45dc-bd5e-e5ae125899e4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.632662] env[61824]: DEBUG nova.compute.manager [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Refreshing instance network info cache due to event network-changed-1fd63140-442e-45dc-bd5e-e5ae125899e4. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 646.632977] env[61824]: DEBUG oslo_concurrency.lockutils [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] Acquiring lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.632977] env[61824]: DEBUG oslo_concurrency.lockutils [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] Acquired lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.633532] env[61824]: DEBUG nova.network.neutron [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Refreshing network info cache for port 1fd63140-442e-45dc-bd5e-e5ae125899e4 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 646.754082] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.754082] env[61824]: DEBUG nova.network.neutron [-] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.152167] env[61824]: DEBUG nova.network.neutron [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.198272] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5807ef7a-15c5-4a0d-a278-d66c9c23db32 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.206873] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79baaf3-c685-40f0-a261-e38b3d8a6571 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.245200] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d88a45c8-5544-4dbd-835d-8e09cfacdd81 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.253221] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b04eb307-2117-45e2-b2d4-de82876f749f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.266043] env[61824]: INFO nova.compute.manager [-] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Took 1.04 seconds to deallocate network for instance. [ 647.273761] env[61824]: DEBUG nova.compute.provider_tree [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.275220] env[61824]: DEBUG nova.compute.claims [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 647.276297] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.359729] env[61824]: DEBUG nova.network.neutron [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.487970] env[61824]: ERROR nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. [ 647.487970] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.487970] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.487970] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.487970] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.487970] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.487970] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.487970] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.487970] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.487970] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 647.487970] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.487970] env[61824]: ERROR nova.compute.manager raise self.value [ 647.487970] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.487970] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.487970] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.487970] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.488636] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.488636] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.488636] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. [ 647.488636] env[61824]: ERROR nova.compute.manager [ 647.488636] env[61824]: Traceback (most recent call last): [ 647.488636] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.488636] env[61824]: listener.cb(fileno) [ 647.488636] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.488636] env[61824]: result = function(*args, **kwargs) [ 647.488636] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.488636] env[61824]: return func(*args, **kwargs) [ 647.488636] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.488636] env[61824]: raise e [ 647.488636] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.488636] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 647.488636] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.488636] env[61824]: created_port_ids = self._update_ports_for_instance( [ 647.488636] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.488636] env[61824]: with excutils.save_and_reraise_exception(): [ 647.488636] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.488636] env[61824]: self.force_reraise() [ 647.488636] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.488636] env[61824]: raise self.value [ 647.488636] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.488636] env[61824]: updated_port = self._update_port( [ 647.488636] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.488636] env[61824]: _ensure_no_port_binding_failure(port) [ 647.488636] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.488636] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.489509] env[61824]: nova.exception.PortBindingFailed: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. [ 647.489509] env[61824]: Removing descriptor: 19 [ 647.776664] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.780752] env[61824]: DEBUG nova.scheduler.client.report [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.806279] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.806601] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.806701] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.806857] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.807009] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.807213] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.807425] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.807584] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.807745] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.807905] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.808090] env[61824]: DEBUG nova.virt.hardware [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.809164] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f49d0b7-ab3d-4835-a869-84e18566ab7e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.817524] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc80b9f-3725-44ef-b74b-99e0a2c7876c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.831799] env[61824]: ERROR nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Traceback (most recent call last): [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] yield resources [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self.driver.spawn(context, instance, image_meta, [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] vm_ref = self.build_virtual_machine(instance, [ 647.831799] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] for vif in network_info: [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] return self._sync_wrapper(fn, *args, **kwargs) [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self.wait() [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self[:] = self._gt.wait() [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] return self._exit_event.wait() [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 647.832167] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] current.throw(*self._exc) [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] result = function(*args, **kwargs) [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] return func(*args, **kwargs) [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] raise e [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] nwinfo = self.network_api.allocate_for_instance( [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] created_port_ids = self._update_ports_for_instance( [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] with excutils.save_and_reraise_exception(): [ 647.832656] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self.force_reraise() [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] raise self.value [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] updated_port = self._update_port( [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] _ensure_no_port_binding_failure(port) [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] raise exception.PortBindingFailed(port_id=port['id']) [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] nova.exception.PortBindingFailed: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. [ 647.833199] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] [ 647.833199] env[61824]: INFO nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Terminating instance [ 647.834105] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Acquiring lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.834263] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Acquired lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.834427] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 647.863286] env[61824]: DEBUG oslo_concurrency.lockutils [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] Releasing lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.863562] env[61824]: DEBUG nova.compute.manager [req-0a88482e-68b5-4d3a-9283-845155ef9a1e req-533a90d4-f4ac-42e8-88d4-5972e3b1e0b0 service nova] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Received event network-vif-deleted-1fd63140-442e-45dc-bd5e-e5ae125899e4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.121247] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "7e73f8b5-a138-4455-a392-9a2b2b860558" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.121247] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.285384] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.285900] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 648.288858] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.729s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.290231] env[61824]: INFO nova.compute.claims [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.354947] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.432642] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.521536] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "7f6a4446-86e1-44c7-ab42-297b033dbace" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.522779] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.698553] env[61824]: DEBUG nova.compute.manager [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Received event network-changed-fec9e102-4433-4cbd-8d36-a1d5dfb250c6 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.698673] env[61824]: DEBUG nova.compute.manager [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Refreshing instance network info cache due to event network-changed-fec9e102-4433-4cbd-8d36-a1d5dfb250c6. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 648.698863] env[61824]: DEBUG oslo_concurrency.lockutils [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] Acquiring lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.802756] env[61824]: DEBUG nova.compute.utils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.808019] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.808019] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 648.863805] env[61824]: DEBUG nova.policy [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87696ef814fe4537b4ef88cc35fb47c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd950efe18e9748fcb5b8b8e48e1e7afd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.934143] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Releasing lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.934587] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.934779] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.935110] env[61824]: DEBUG oslo_concurrency.lockutils [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] Acquired lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.935285] env[61824]: DEBUG nova.network.neutron [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Refreshing network info cache for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 648.936432] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8134ffd-8847-4b87-8c41-a04fae334452 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.948728] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-584e630e-fa9c-48b4-a05f-788bf74e481b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.970110] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d could not be found. [ 648.970234] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.970495] env[61824]: INFO nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 648.970708] env[61824]: DEBUG oslo.service.loopingcall [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.970922] env[61824]: DEBUG nova.compute.manager [-] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.971027] env[61824]: DEBUG nova.network.neutron [-] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.985662] env[61824]: DEBUG nova.network.neutron [-] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.135670] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Successfully created port: 1f433b02-2760-4410-9da4-72563f09ae3e {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.307177] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.456419] env[61824]: DEBUG nova.network.neutron [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.488566] env[61824]: DEBUG nova.network.neutron [-] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.547900] env[61824]: DEBUG nova.network.neutron [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.721112] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008fea8c-f731-4ad5-b7b3-375c94f3f943 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.729358] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242655af-d3f9-4144-93c4-1018b2e908f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.767156] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38fd32c6-c80b-4c98-b988-e12e2206ecb3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.774688] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ab5721-0c33-46dc-8986-a618f2eca79e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.788638] env[61824]: DEBUG nova.compute.provider_tree [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.997515] env[61824]: INFO nova.compute.manager [-] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Took 1.03 seconds to deallocate network for instance. [ 650.000739] env[61824]: DEBUG nova.compute.claims [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 650.000739] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.051478] env[61824]: DEBUG oslo_concurrency.lockutils [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] Releasing lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.051478] env[61824]: DEBUG nova.compute.manager [req-ffb254f6-79f9-46a9-9f14-cd82bedbe536 req-8c3fc9f6-9e52-4958-8e7c-54ccfa258153 service nova] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Received event network-vif-deleted-fec9e102-4433-4cbd-8d36-a1d5dfb250c6 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.102401] env[61824]: ERROR nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. [ 650.102401] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 650.102401] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.102401] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 650.102401] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.102401] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 650.102401] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.102401] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 650.102401] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.102401] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 650.102401] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.102401] env[61824]: ERROR nova.compute.manager raise self.value [ 650.102401] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.102401] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 650.102401] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.102401] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 650.103031] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.103031] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 650.103031] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. [ 650.103031] env[61824]: ERROR nova.compute.manager [ 650.103031] env[61824]: Traceback (most recent call last): [ 650.103031] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 650.103031] env[61824]: listener.cb(fileno) [ 650.103031] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.103031] env[61824]: result = function(*args, **kwargs) [ 650.103031] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.103031] env[61824]: return func(*args, **kwargs) [ 650.103031] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.103031] env[61824]: raise e [ 650.103031] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.103031] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 650.103031] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.103031] env[61824]: created_port_ids = self._update_ports_for_instance( [ 650.103031] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.103031] env[61824]: with excutils.save_and_reraise_exception(): [ 650.103031] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.103031] env[61824]: self.force_reraise() [ 650.103031] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.103031] env[61824]: raise self.value [ 650.103031] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.103031] env[61824]: updated_port = self._update_port( [ 650.103031] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.103031] env[61824]: _ensure_no_port_binding_failure(port) [ 650.103031] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.103031] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 650.103808] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. [ 650.103808] env[61824]: Removing descriptor: 19 [ 650.291445] env[61824]: DEBUG nova.scheduler.client.report [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.314254] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.338632] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.339083] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.339083] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.339217] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.339363] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.339598] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.339709] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.339862] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.340036] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.340202] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.340375] env[61824]: DEBUG nova.virt.hardware [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.341243] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da132f81-afb0-4f17-b10b-37b556a367c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.349845] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379b0414-93a0-48b4-ab4f-7d8590d304e7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.363617] env[61824]: ERROR nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Traceback (most recent call last): [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] yield resources [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self.driver.spawn(context, instance, image_meta, [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] vm_ref = self.build_virtual_machine(instance, [ 650.363617] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] for vif in network_info: [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] return self._sync_wrapper(fn, *args, **kwargs) [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self.wait() [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self[:] = self._gt.wait() [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] return self._exit_event.wait() [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 650.363970] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] current.throw(*self._exc) [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] result = function(*args, **kwargs) [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] return func(*args, **kwargs) [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] raise e [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] nwinfo = self.network_api.allocate_for_instance( [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] created_port_ids = self._update_ports_for_instance( [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] with excutils.save_and_reraise_exception(): [ 650.364385] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self.force_reraise() [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] raise self.value [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] updated_port = self._update_port( [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] _ensure_no_port_binding_failure(port) [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] raise exception.PortBindingFailed(port_id=port['id']) [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] nova.exception.PortBindingFailed: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. [ 650.364885] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] [ 650.364885] env[61824]: INFO nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Terminating instance [ 650.365932] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Acquiring lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.366157] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Acquired lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.366338] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.753216] env[61824]: DEBUG nova.compute.manager [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Received event network-changed-1f433b02-2760-4410-9da4-72563f09ae3e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.753424] env[61824]: DEBUG nova.compute.manager [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Refreshing instance network info cache due to event network-changed-1f433b02-2760-4410-9da4-72563f09ae3e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 650.753614] env[61824]: DEBUG oslo_concurrency.lockutils [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] Acquiring lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.796460] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.797168] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.800732] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.610s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.885464] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.967951] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.304333] env[61824]: DEBUG nova.compute.utils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 651.310209] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 651.310331] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 651.350485] env[61824]: DEBUG nova.policy [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6a2f1ffe8934519acdcaca77e2a55d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e51dd2f578344aa8afa5d6beccb54f1e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.471675] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Releasing lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.472122] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 651.472315] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 651.472623] env[61824]: DEBUG oslo_concurrency.lockutils [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] Acquired lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.472796] env[61824]: DEBUG nova.network.neutron [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Refreshing network info cache for port 1f433b02-2760-4410-9da4-72563f09ae3e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 651.473883] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd8a5873-f6bb-4039-8528-c9578682b1e4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.482917] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c7d48c-99d6-47d8-a7b3-363cf8b14abc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.506870] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3cd00254-b61b-498a-946b-ff2028230e25 could not be found. [ 651.506870] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.506870] env[61824]: INFO nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Took 0.03 seconds to destroy the instance on the hypervisor. [ 651.506870] env[61824]: DEBUG oslo.service.loopingcall [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.506870] env[61824]: DEBUG nova.compute.manager [-] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.506870] env[61824]: DEBUG nova.network.neutron [-] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.529724] env[61824]: DEBUG nova.network.neutron [-] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.600684] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Successfully created port: 3d8f0991-2a49-4263-8cb1-0197df211305 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.810790] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.839450] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 95635ec3-cbfd-4dc0-a595-02062a0bda6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.839450] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 496f00ef-7487-45d9-9356-529f9390544f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.839450] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance ad02e0c7-3d12-48ad-a393-26749fd8c502 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.839450] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance d51370a2-86d5-4c9e-978c-f3318acfd3a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.839668] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 325a641d-2509-49e6-bfbc-0b2cb85c1ff9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.839668] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.839668] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 3cd00254-b61b-498a-946b-ff2028230e25 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.839668] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance e56a6950-f842-45fb-a0a0-0defe3b3c30f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 651.993712] env[61824]: DEBUG nova.network.neutron [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.033250] env[61824]: DEBUG nova.network.neutron [-] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.103792] env[61824]: DEBUG nova.network.neutron [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.343911] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 9ce52867-d547-401e-993d-0053eb9da4c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.439872] env[61824]: ERROR nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. [ 652.439872] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.439872] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.439872] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.439872] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.439872] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.439872] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.439872] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.439872] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.439872] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 652.439872] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.439872] env[61824]: ERROR nova.compute.manager raise self.value [ 652.439872] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.439872] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.439872] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.439872] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.440326] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.440326] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.440326] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. [ 652.440326] env[61824]: ERROR nova.compute.manager [ 652.440326] env[61824]: Traceback (most recent call last): [ 652.440326] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.440326] env[61824]: listener.cb(fileno) [ 652.440326] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.440326] env[61824]: result = function(*args, **kwargs) [ 652.440326] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.440326] env[61824]: return func(*args, **kwargs) [ 652.440326] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.440326] env[61824]: raise e [ 652.440326] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.440326] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 652.440326] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.440326] env[61824]: created_port_ids = self._update_ports_for_instance( [ 652.440326] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.440326] env[61824]: with excutils.save_and_reraise_exception(): [ 652.440326] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.440326] env[61824]: self.force_reraise() [ 652.440326] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.440326] env[61824]: raise self.value [ 652.440326] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.440326] env[61824]: updated_port = self._update_port( [ 652.440326] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.440326] env[61824]: _ensure_no_port_binding_failure(port) [ 652.440326] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.440326] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.441049] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. [ 652.441049] env[61824]: Removing descriptor: 19 [ 652.539939] env[61824]: INFO nova.compute.manager [-] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Took 1.03 seconds to deallocate network for instance. [ 652.542337] env[61824]: DEBUG nova.compute.claims [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 652.542557] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.606395] env[61824]: DEBUG oslo_concurrency.lockutils [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] Releasing lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.606679] env[61824]: DEBUG nova.compute.manager [req-2925930a-f4d8-4037-abd9-8c9e36a255d1 req-d5cdb1f6-bfc6-43a6-a11a-b169861f2eae service nova] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Received event network-vif-deleted-1f433b02-2760-4410-9da4-72563f09ae3e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.781789] env[61824]: DEBUG nova.compute.manager [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Received event network-changed-3d8f0991-2a49-4263-8cb1-0197df211305 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.781949] env[61824]: DEBUG nova.compute.manager [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Refreshing instance network info cache due to event network-changed-3d8f0991-2a49-4263-8cb1-0197df211305. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 652.782218] env[61824]: DEBUG oslo_concurrency.lockutils [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] Acquiring lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.782311] env[61824]: DEBUG oslo_concurrency.lockutils [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] Acquired lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.782631] env[61824]: DEBUG nova.network.neutron [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Refreshing network info cache for port 3d8f0991-2a49-4263-8cb1-0197df211305 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 652.823510] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.847708] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.847708] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.847890] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.847949] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.848134] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.848286] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.848491] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.848653] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.848817] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.848980] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.849175] env[61824]: DEBUG nova.virt.hardware [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.849874] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7a93d1c2-20d6-438e-bc42-1f9aa0219f91 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 652.851909] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f885cf-950f-4ba3-8fea-9ca8fb455b0f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.861902] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e86a991-ff40-4013-b3a7-6207dc09efee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.877438] env[61824]: ERROR nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Traceback (most recent call last): [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] yield resources [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self.driver.spawn(context, instance, image_meta, [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] vm_ref = self.build_virtual_machine(instance, [ 652.877438] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] for vif in network_info: [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] return self._sync_wrapper(fn, *args, **kwargs) [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self.wait() [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self[:] = self._gt.wait() [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] return self._exit_event.wait() [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 652.878063] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] current.throw(*self._exc) [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] result = function(*args, **kwargs) [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] return func(*args, **kwargs) [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] raise e [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] nwinfo = self.network_api.allocate_for_instance( [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] created_port_ids = self._update_ports_for_instance( [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] with excutils.save_and_reraise_exception(): [ 652.878447] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self.force_reraise() [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] raise self.value [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] updated_port = self._update_port( [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] _ensure_no_port_binding_failure(port) [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] raise exception.PortBindingFailed(port_id=port['id']) [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] nova.exception.PortBindingFailed: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. [ 652.878757] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] [ 652.878757] env[61824]: INFO nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Terminating instance [ 652.882120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.302073] env[61824]: DEBUG nova.network.neutron [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.356676] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 717eca7a-0287-4424-a302-5737f92724a6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.384045] env[61824]: DEBUG nova.network.neutron [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.859731] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8aeeb8e9-e1ab-463b-bc16-ca68571882aa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 653.888788] env[61824]: DEBUG oslo_concurrency.lockutils [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] Releasing lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.889085] env[61824]: DEBUG nova.compute.manager [req-d88887d9-30c2-4ad9-9b3a-794b095efd0a req-d752ab37-9813-4a04-bd46-71a8d561c6c1 service nova] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Received event network-vif-deleted-3d8f0991-2a49-4263-8cb1-0197df211305 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.889808] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.889808] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.363478] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.406294] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.476088] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.866467] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 037dd0ba-0d50-4c46-8947-7df3ca19e1c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 654.978587] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.979086] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 654.979324] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 654.979656] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4ca853b7-d07d-4790-8d6c-1e6fd73f2879 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.988948] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a40d0a-ddaa-40ca-a4e7-574aa538cdd2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.009805] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e56a6950-f842-45fb-a0a0-0defe3b3c30f could not be found. [ 655.010012] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.010201] env[61824]: INFO nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 655.010459] env[61824]: DEBUG oslo.service.loopingcall [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.010677] env[61824]: DEBUG nova.compute.manager [-] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.010768] env[61824]: DEBUG nova.network.neutron [-] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.025016] env[61824]: DEBUG nova.network.neutron [-] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.369581] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 6e2190ac-d287-4810-9d43-740e62ae7b56 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 655.528096] env[61824]: DEBUG nova.network.neutron [-] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.872891] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance dc662f50-f111-4d26-b017-3ab719004ac2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.030451] env[61824]: INFO nova.compute.manager [-] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Took 1.02 seconds to deallocate network for instance. [ 656.033616] env[61824]: DEBUG nova.compute.claims [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.033847] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.375868] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 662e815c-2548-4aed-a928-d98f9e1297e7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 656.879173] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance e758c89f-e5e2-42e0-926e-c629cb496a89 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.383611] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 204fd4be-4a16-4838-a30d-c78339b3834d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 657.887031] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance f525feb8-d24a-4f89-b640-471163d181ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.389902] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 9942dd4b-8c04-40ac-b9f4-318b13902535 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 658.893757] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance d889d359-f767-4853-8092-070e7656c284 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.397473] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 61d1f216-7ff1-419d-b415-3dce3537b1b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 659.900826] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 329b3d0d-7007-40e3-a554-fbd0c7497b69 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.404315] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 3f89fb2d-99c1-4cf7-83aa-db033211a2ce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 660.907548] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 9fd8e236-3556-4b95-952f-3c324b896e29 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.410548] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 34f72773-5950-4710-b3e8-5e5c4877f227 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.913862] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 0f86f95d-2c41-47b2-93f8-c7e6983c913a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.416697] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7e73f8b5-a138-4455-a392-9a2b2b860558 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.919994] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7f6a4446-86e1-44c7-ab42-297b033dbace has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.920296] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 662.920452] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 663.314755] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9b8a1b-15d2-4eda-bbcf-59280fd9e180 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.322803] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772eda3f-aa2c-4f8c-a62b-ca2dc5af086c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.352268] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3b0996-18ae-4a87-bda0-0e35da163278 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.359552] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82725a9a-4987-4d84-bc39-4edb973cafa6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.372978] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.876062] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.382772] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 664.382772] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.581s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.382772] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.308s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.208550] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83f4ecf-f661-42d7-9b9e-ac6b6eb78a39 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.216250] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e11713-2ffc-4c31-b69f-bbcfaa8e6620 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.247796] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0672bde-21dc-4c62-b91c-8e1a970ddb00 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.256147] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febc4a4d-6d84-40fd-866c-d4d26f3628a0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.269605] env[61824]: DEBUG nova.compute.provider_tree [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.773016] env[61824]: DEBUG nova.scheduler.client.report [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.277387] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.895s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.278138] env[61824]: ERROR nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Traceback (most recent call last): [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self.driver.spawn(context, instance, image_meta, [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] vm_ref = self.build_virtual_machine(instance, [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.278138] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] for vif in network_info: [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return self._sync_wrapper(fn, *args, **kwargs) [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self.wait() [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self[:] = self._gt.wait() [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return self._exit_event.wait() [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] result = hub.switch() [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.278434] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return self.greenlet.switch() [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] result = function(*args, **kwargs) [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] return func(*args, **kwargs) [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] raise e [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] nwinfo = self.network_api.allocate_for_instance( [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] created_port_ids = self._update_ports_for_instance( [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] with excutils.save_and_reraise_exception(): [ 666.278775] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] self.force_reraise() [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] raise self.value [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] updated_port = self._update_port( [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] _ensure_no_port_binding_failure(port) [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] raise exception.PortBindingFailed(port_id=port['id']) [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] nova.exception.PortBindingFailed: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. [ 666.279101] env[61824]: ERROR nova.compute.manager [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] [ 666.279455] env[61824]: DEBUG nova.compute.utils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 666.280072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.171s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.283200] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Build of instance 95635ec3-cbfd-4dc0-a595-02062a0bda6e was re-scheduled: Binding failed for port 138652a9-ac69-46e0-ab61-047a785a8e9b, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 666.283619] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 666.283891] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Acquiring lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.284057] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Acquired lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.284222] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 666.805253] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.850172] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.121389] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627f15e0-2b34-4d6b-acce-e914df93cb06 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.129891] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def92d99-f159-4be6-a860-b970e1e4b1b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.161745] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90c39b4-4fc8-4a01-9d4d-bc1439b80ae2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.169258] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c802ba0-8109-474d-8e97-0210d619d98c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.182812] env[61824]: DEBUG nova.compute.provider_tree [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.352328] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Releasing lock "refresh_cache-95635ec3-cbfd-4dc0-a595-02062a0bda6e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.352606] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 667.352742] env[61824]: DEBUG nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.352894] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 667.369933] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.686033] env[61824]: DEBUG nova.scheduler.client.report [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 667.872048] env[61824]: DEBUG nova.network.neutron [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.190873] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.191560] env[61824]: ERROR nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] Traceback (most recent call last): [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self.driver.spawn(context, instance, image_meta, [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] vm_ref = self.build_virtual_machine(instance, [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.191560] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] for vif in network_info: [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return self._sync_wrapper(fn, *args, **kwargs) [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self.wait() [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self[:] = self._gt.wait() [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return self._exit_event.wait() [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] result = hub.switch() [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.191865] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return self.greenlet.switch() [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] result = function(*args, **kwargs) [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] return func(*args, **kwargs) [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] raise e [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] nwinfo = self.network_api.allocate_for_instance( [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] created_port_ids = self._update_ports_for_instance( [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] with excutils.save_and_reraise_exception(): [ 668.192281] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] self.force_reraise() [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] raise self.value [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] updated_port = self._update_port( [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] _ensure_no_port_binding_failure(port) [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] raise exception.PortBindingFailed(port_id=port['id']) [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] nova.exception.PortBindingFailed: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. [ 668.192604] env[61824]: ERROR nova.compute.manager [instance: 496f00ef-7487-45d9-9356-529f9390544f] [ 668.192870] env[61824]: DEBUG nova.compute.utils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.193516] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.801s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.197609] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Build of instance 496f00ef-7487-45d9-9356-529f9390544f was re-scheduled: Binding failed for port 09195547-1b85-432a-8167-473f4d76bd77, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.197609] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.197609] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Acquiring lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.197609] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Acquired lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.197855] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.375193] env[61824]: INFO nova.compute.manager [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] [instance: 95635ec3-cbfd-4dc0-a595-02062a0bda6e] Took 1.02 seconds to deallocate network for instance. [ 668.719184] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.767833] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.037144] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323121d6-4444-49c5-b6fd-9411299adbeb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.044849] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af1ad74-136a-4e49-a170-c28c7d8ed155 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.076459] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e23c8a5-911a-47a3-94aa-0a34168b6902 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.083774] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68bbf56-1694-4caa-a6a2-dfda815fb651 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.096962] env[61824]: DEBUG nova.compute.provider_tree [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.272590] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Releasing lock "refresh_cache-496f00ef-7487-45d9-9356-529f9390544f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.272839] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 669.273508] env[61824]: DEBUG nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.273508] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.292087] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.401441] env[61824]: INFO nova.scheduler.client.report [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Deleted allocations for instance 95635ec3-cbfd-4dc0-a595-02062a0bda6e [ 669.600327] env[61824]: DEBUG nova.scheduler.client.report [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.794907] env[61824]: DEBUG nova.network.neutron [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.913159] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5b941055-1716-4835-a37d-3c72852930b4 tempest-TenantUsagesTestJSON-1948251512 tempest-TenantUsagesTestJSON-1948251512-project-member] Lock "95635ec3-cbfd-4dc0-a595-02062a0bda6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.636s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.106018] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.106808] env[61824]: ERROR nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Traceback (most recent call last): [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self.driver.spawn(context, instance, image_meta, [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] vm_ref = self.build_virtual_machine(instance, [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.106808] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] for vif in network_info: [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] return self._sync_wrapper(fn, *args, **kwargs) [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self.wait() [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self[:] = self._gt.wait() [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] return self._exit_event.wait() [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] current.throw(*self._exc) [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.107198] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] result = function(*args, **kwargs) [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] return func(*args, **kwargs) [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] raise e [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] nwinfo = self.network_api.allocate_for_instance( [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] created_port_ids = self._update_ports_for_instance( [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] with excutils.save_and_reraise_exception(): [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] self.force_reraise() [ 670.107549] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] raise self.value [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] updated_port = self._update_port( [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] _ensure_no_port_binding_failure(port) [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] raise exception.PortBindingFailed(port_id=port['id']) [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] nova.exception.PortBindingFailed: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. [ 670.107863] env[61824]: ERROR nova.compute.manager [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] [ 670.107863] env[61824]: DEBUG nova.compute.utils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.108942] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.373s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.110527] env[61824]: INFO nova.compute.claims [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.113356] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Build of instance ad02e0c7-3d12-48ad-a393-26749fd8c502 was re-scheduled: Binding failed for port 5910859c-229c-4d45-b859-61b6bafce7e8, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.114013] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 670.114363] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Acquiring lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.114607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Acquired lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.114856] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 670.299715] env[61824]: INFO nova.compute.manager [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] [instance: 496f00ef-7487-45d9-9356-529f9390544f] Took 1.03 seconds to deallocate network for instance. [ 670.415464] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 670.637965] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.736083] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.944079] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.239361] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Releasing lock "refresh_cache-ad02e0c7-3d12-48ad-a393-26749fd8c502" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.239600] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 671.239784] env[61824]: DEBUG nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.240008] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 671.256899] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.327217] env[61824]: INFO nova.scheduler.client.report [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Deleted allocations for instance 496f00ef-7487-45d9-9356-529f9390544f [ 671.477942] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf17ea8-fa41-4931-a009-982890e17b21 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.486142] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c54d04-60db-48d6-9095-bdfe68700f1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.515172] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae63f82e-be7a-4399-aa89-b3c8f31551b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.522179] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0c79c9-6e46-4236-be6a-9ab17f3d9ae2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.535717] env[61824]: DEBUG nova.compute.provider_tree [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.761615] env[61824]: DEBUG nova.network.neutron [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.836567] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8235310e-fd4a-4cca-8e13-0721d5d5800d tempest-ServerDiagnosticsNegativeTest-472065821 tempest-ServerDiagnosticsNegativeTest-472065821-project-member] Lock "496f00ef-7487-45d9-9356-529f9390544f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.469s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.040524] env[61824]: DEBUG nova.scheduler.client.report [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.264795] env[61824]: INFO nova.compute.manager [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] [instance: ad02e0c7-3d12-48ad-a393-26749fd8c502] Took 1.02 seconds to deallocate network for instance. [ 672.339141] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 672.546479] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.547007] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 672.549944] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.865s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.551729] env[61824]: INFO nova.compute.claims [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.864158] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.056149] env[61824]: DEBUG nova.compute.utils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 673.059566] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 673.059765] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 673.128118] env[61824]: DEBUG nova.policy [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '161a76e53a1045d3a14b7b9dac2f1c2c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cf6af3ecef57440dbf33ebddd3ea25de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.303811] env[61824]: INFO nova.scheduler.client.report [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Deleted allocations for instance ad02e0c7-3d12-48ad-a393-26749fd8c502 [ 673.438186] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Successfully created port: 35ea00d6-de28-4d0c-9fcb-92fca8a21068 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.560727] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 673.818506] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e513e4e0-cc0f-414d-bb2e-31445ee94458 tempest-AttachInterfacesUnderV243Test-1460435660 tempest-AttachInterfacesUnderV243Test-1460435660-project-member] Lock "ad02e0c7-3d12-48ad-a393-26749fd8c502" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.173s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.044334] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af5cf4a-7472-4613-99a8-6e3eeb9b3335 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.053463] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-028c0dc8-065e-4368-a059-4cd044f4161a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.092754] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ccece3-bd00-404f-8f29-3d4ab42dbf9d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.103579] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6eaabe-a739-4ae1-b7c9-9f03b01debb3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.124816] env[61824]: DEBUG nova.compute.provider_tree [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.313377] env[61824]: DEBUG nova.compute.manager [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Received event network-changed-35ea00d6-de28-4d0c-9fcb-92fca8a21068 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.313592] env[61824]: DEBUG nova.compute.manager [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Refreshing instance network info cache due to event network-changed-35ea00d6-de28-4d0c-9fcb-92fca8a21068. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 674.313781] env[61824]: DEBUG oslo_concurrency.lockutils [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] Acquiring lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.313920] env[61824]: DEBUG oslo_concurrency.lockutils [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] Acquired lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.314086] env[61824]: DEBUG nova.network.neutron [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Refreshing network info cache for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 674.324458] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 674.557899] env[61824]: ERROR nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. [ 674.557899] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.557899] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.557899] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.557899] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.557899] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.557899] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.557899] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.557899] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.557899] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 674.557899] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.557899] env[61824]: ERROR nova.compute.manager raise self.value [ 674.557899] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.557899] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.557899] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.557899] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.558530] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.558530] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.558530] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. [ 674.558530] env[61824]: ERROR nova.compute.manager [ 674.558530] env[61824]: Traceback (most recent call last): [ 674.558530] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.558530] env[61824]: listener.cb(fileno) [ 674.558530] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.558530] env[61824]: result = function(*args, **kwargs) [ 674.558530] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.558530] env[61824]: return func(*args, **kwargs) [ 674.558530] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.558530] env[61824]: raise e [ 674.558530] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.558530] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 674.558530] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.558530] env[61824]: created_port_ids = self._update_ports_for_instance( [ 674.558530] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.558530] env[61824]: with excutils.save_and_reraise_exception(): [ 674.558530] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.558530] env[61824]: self.force_reraise() [ 674.558530] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.558530] env[61824]: raise self.value [ 674.558530] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.558530] env[61824]: updated_port = self._update_port( [ 674.558530] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.558530] env[61824]: _ensure_no_port_binding_failure(port) [ 674.558530] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.558530] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.559416] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. [ 674.559416] env[61824]: Removing descriptor: 19 [ 674.593413] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.631207] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.631207] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.631207] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.631416] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.631416] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.631416] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.631416] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.631416] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.631571] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.631571] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.631571] env[61824]: DEBUG nova.virt.hardware [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.631752] env[61824]: DEBUG nova.scheduler.client.report [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 674.635729] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d34363f-ac4a-408c-9db9-d9dc31fbaf62 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.651284] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e884c9c-48e3-4a08-911f-3594818192b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.668532] env[61824]: ERROR nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Traceback (most recent call last): [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] yield resources [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self.driver.spawn(context, instance, image_meta, [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] vm_ref = self.build_virtual_machine(instance, [ 674.668532] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] for vif in network_info: [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] return self._sync_wrapper(fn, *args, **kwargs) [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self.wait() [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self[:] = self._gt.wait() [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] return self._exit_event.wait() [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.671358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] current.throw(*self._exc) [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] result = function(*args, **kwargs) [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] return func(*args, **kwargs) [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] raise e [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] nwinfo = self.network_api.allocate_for_instance( [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] created_port_ids = self._update_ports_for_instance( [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] with excutils.save_and_reraise_exception(): [ 674.671682] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self.force_reraise() [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] raise self.value [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] updated_port = self._update_port( [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] _ensure_no_port_binding_failure(port) [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] raise exception.PortBindingFailed(port_id=port['id']) [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] nova.exception.PortBindingFailed: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. [ 674.672030] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] [ 674.672030] env[61824]: INFO nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Terminating instance [ 674.672347] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Acquiring lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.835869] env[61824]: DEBUG nova.network.neutron [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.851336] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.963192] env[61824]: DEBUG nova.network.neutron [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.142829] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.143379] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.146407] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.068s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.470247] env[61824]: DEBUG oslo_concurrency.lockutils [req-38b86da9-2471-4e03-8b46-385eada4af2a req-b11e6918-b976-480d-942e-722fea438740 service nova] Releasing lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.470848] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Acquired lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.471155] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 675.653288] env[61824]: DEBUG nova.compute.utils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.657485] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.657918] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 675.753738] env[61824]: DEBUG nova.policy [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '632dafc515e34f318f6f4b43bd1a62aa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '783bb630c83943f286b3d019c2832ba1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 675.995593] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.099880] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cff0c9-a39e-4c48-9b8e-1d8210b2c138 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.110854] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f131e1e-76dc-4ca4-97a0-95c68ab03e4c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.147170] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Successfully created port: 33a84cb8-47fb-42c6-b23f-6cd67baba2ea {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.149824] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.155208] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a226d4-b3a3-4463-b270-61ff6ae37276 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.160731] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.165979] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d517bace-3568-4e37-8d4f-36ad7fa7a6df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.183821] env[61824]: DEBUG nova.compute.provider_tree [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.357070] env[61824]: DEBUG nova.compute.manager [req-4cf33e52-a536-42a5-ae5b-2945fd9b07d0 req-70c66dc1-8ab6-49f8-87b7-1431f612d95d service nova] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Received event network-vif-deleted-35ea00d6-de28-4d0c-9fcb-92fca8a21068 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 676.658009] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Releasing lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.658459] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 676.658651] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 676.659103] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1e73c013-9385-46a9-9efd-183816b1cfd1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.671285] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6b8f18-f25e-4f77-98ea-e29a4ad1ab1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.693233] env[61824]: DEBUG nova.scheduler.client.report [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.696770] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ce52867-d547-401e-993d-0053eb9da4c4 could not be found. [ 676.696942] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.700845] env[61824]: INFO nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 676.700845] env[61824]: DEBUG oslo.service.loopingcall [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 676.701695] env[61824]: DEBUG nova.compute.manager [-] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.701984] env[61824]: DEBUG nova.network.neutron [-] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 676.745918] env[61824]: DEBUG nova.network.neutron [-] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.183420] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.202683] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.056s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.203411] env[61824]: ERROR nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Traceback (most recent call last): [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self.driver.spawn(context, instance, image_meta, [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] vm_ref = self.build_virtual_machine(instance, [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.203411] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] for vif in network_info: [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return self._sync_wrapper(fn, *args, **kwargs) [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self.wait() [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self[:] = self._gt.wait() [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return self._exit_event.wait() [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] result = hub.switch() [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.203738] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return self.greenlet.switch() [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] result = function(*args, **kwargs) [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] return func(*args, **kwargs) [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] raise e [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] nwinfo = self.network_api.allocate_for_instance( [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] created_port_ids = self._update_ports_for_instance( [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] with excutils.save_and_reraise_exception(): [ 677.204089] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] self.force_reraise() [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] raise self.value [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] updated_port = self._update_port( [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] _ensure_no_port_binding_failure(port) [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] raise exception.PortBindingFailed(port_id=port['id']) [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] nova.exception.PortBindingFailed: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. [ 677.204402] env[61824]: ERROR nova.compute.manager [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] [ 677.204667] env[61824]: DEBUG nova.compute.utils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.206113] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.930s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.212287] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Build of instance d51370a2-86d5-4c9e-978c-f3318acfd3a7 was re-scheduled: Binding failed for port 1a7fc60a-1f66-40ad-b353-e5d0a2d77403, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.212287] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.212287] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.212287] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.212558] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.213362] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.213628] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.213846] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.214473] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.214473] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.214473] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.214708] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.215037] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.215270] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.215485] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.215703] env[61824]: DEBUG nova.virt.hardware [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.220019] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af5d9dd-1880-40e6-9757-80a256471794 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.227865] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467e45ae-1d49-4c12-8f01-77e81b4c3ce0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.250303] env[61824]: DEBUG nova.network.neutron [-] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.297349] env[61824]: ERROR nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. [ 677.297349] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.297349] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.297349] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.297349] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.297349] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.297349] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.297349] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.297349] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.297349] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 677.297349] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.297349] env[61824]: ERROR nova.compute.manager raise self.value [ 677.297349] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.297349] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.297349] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.297349] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.297776] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.297776] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.297776] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. [ 677.297776] env[61824]: ERROR nova.compute.manager [ 677.297776] env[61824]: Traceback (most recent call last): [ 677.297776] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.297776] env[61824]: listener.cb(fileno) [ 677.297776] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.297776] env[61824]: result = function(*args, **kwargs) [ 677.297776] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.297776] env[61824]: return func(*args, **kwargs) [ 677.297776] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.297776] env[61824]: raise e [ 677.297776] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.297776] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 677.297776] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.297776] env[61824]: created_port_ids = self._update_ports_for_instance( [ 677.297776] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.297776] env[61824]: with excutils.save_and_reraise_exception(): [ 677.297776] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.297776] env[61824]: self.force_reraise() [ 677.297776] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.297776] env[61824]: raise self.value [ 677.297776] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.297776] env[61824]: updated_port = self._update_port( [ 677.297776] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.297776] env[61824]: _ensure_no_port_binding_failure(port) [ 677.297776] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.297776] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.298475] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. [ 677.298475] env[61824]: Removing descriptor: 19 [ 677.298475] env[61824]: ERROR nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Traceback (most recent call last): [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] yield resources [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self.driver.spawn(context, instance, image_meta, [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.298475] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] vm_ref = self.build_virtual_machine(instance, [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] for vif in network_info: [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return self._sync_wrapper(fn, *args, **kwargs) [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self.wait() [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self[:] = self._gt.wait() [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return self._exit_event.wait() [ 677.298783] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] result = hub.switch() [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return self.greenlet.switch() [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] result = function(*args, **kwargs) [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return func(*args, **kwargs) [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] raise e [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] nwinfo = self.network_api.allocate_for_instance( [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 677.299185] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] created_port_ids = self._update_ports_for_instance( [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] with excutils.save_and_reraise_exception(): [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self.force_reraise() [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] raise self.value [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] updated_port = self._update_port( [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] _ensure_no_port_binding_failure(port) [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.299574] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] raise exception.PortBindingFailed(port_id=port['id']) [ 677.299863] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] nova.exception.PortBindingFailed: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. [ 677.299863] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] [ 677.299863] env[61824]: INFO nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Terminating instance [ 677.301470] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.301717] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquired lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.302022] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 677.376489] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.603574] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.754365] env[61824]: INFO nova.compute.manager [-] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Took 1.05 seconds to deallocate network for instance. [ 677.758094] env[61824]: DEBUG nova.compute.claims [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 677.758284] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.827628] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.940082] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.068770] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99695b71-c756-4e2a-9510-691969808548 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.076215] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66d8781-9f3e-4bbc-a249-172e883905ba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.106437] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07aa433-7a1b-4ccd-b160-0a0d9ce0fbbe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.108374] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "refresh_cache-d51370a2-86d5-4c9e-978c-f3318acfd3a7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.108512] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.108693] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.108853] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.115048] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3537cff7-aafc-4295-ba56-3542efb0a2fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.129015] env[61824]: DEBUG nova.compute.provider_tree [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.130289] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.384111] env[61824]: DEBUG nova.compute.manager [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Received event network-changed-33a84cb8-47fb-42c6-b23f-6cd67baba2ea {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.384111] env[61824]: DEBUG nova.compute.manager [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Refreshing instance network info cache due to event network-changed-33a84cb8-47fb-42c6-b23f-6cd67baba2ea. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 678.384111] env[61824]: DEBUG oslo_concurrency.lockutils [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] Acquiring lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.450019] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Releasing lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.450019] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 678.450019] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 678.450019] env[61824]: DEBUG oslo_concurrency.lockutils [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] Acquired lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.450019] env[61824]: DEBUG nova.network.neutron [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Refreshing network info cache for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 678.450884] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc27dec9-f0d6-416e-8929-8623e5b7accf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.461548] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4391fadf-a5c0-4ae6-ba27-801591058c3e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.489140] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7a93d1c2-20d6-438e-bc42-1f9aa0219f91 could not be found. [ 678.489140] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.489140] env[61824]: INFO nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Took 0.04 seconds to destroy the instance on the hypervisor. [ 678.489301] env[61824]: DEBUG oslo.service.loopingcall [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.489388] env[61824]: DEBUG nova.compute.manager [-] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.489490] env[61824]: DEBUG nova.network.neutron [-] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 678.513734] env[61824]: DEBUG nova.network.neutron [-] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.636019] env[61824]: DEBUG nova.scheduler.client.report [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.641317] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.977073] env[61824]: DEBUG nova.network.neutron [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.016496] env[61824]: DEBUG nova.network.neutron [-] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.142650] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.142930] env[61824]: ERROR nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Traceback (most recent call last): [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self.driver.spawn(context, instance, image_meta, [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] vm_ref = self.build_virtual_machine(instance, [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.142930] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] for vif in network_info: [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return self._sync_wrapper(fn, *args, **kwargs) [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self.wait() [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self[:] = self._gt.wait() [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return self._exit_event.wait() [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] result = hub.switch() [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 679.143274] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return self.greenlet.switch() [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] result = function(*args, **kwargs) [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] return func(*args, **kwargs) [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] raise e [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] nwinfo = self.network_api.allocate_for_instance( [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] created_port_ids = self._update_ports_for_instance( [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] with excutils.save_and_reraise_exception(): [ 679.143596] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] self.force_reraise() [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] raise self.value [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] updated_port = self._update_port( [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] _ensure_no_port_binding_failure(port) [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] raise exception.PortBindingFailed(port_id=port['id']) [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] nova.exception.PortBindingFailed: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. [ 679.146772] env[61824]: ERROR nova.compute.manager [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] [ 679.147137] env[61824]: DEBUG nova.compute.utils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.147137] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.145s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.148594] env[61824]: INFO nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: d51370a2-86d5-4c9e-978c-f3318acfd3a7] Took 1.04 seconds to deallocate network for instance. [ 679.150600] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Build of instance 325a641d-2509-49e6-bfbc-0b2cb85c1ff9 was re-scheduled: Binding failed for port 1fd63140-442e-45dc-bd5e-e5ae125899e4, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.151050] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.151281] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.151449] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.151580] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 679.161786] env[61824]: DEBUG nova.network.neutron [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.519610] env[61824]: INFO nova.compute.manager [-] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Took 1.03 seconds to deallocate network for instance. [ 679.529031] env[61824]: DEBUG nova.compute.claims [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 679.529031] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.666330] env[61824]: DEBUG oslo_concurrency.lockutils [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] Releasing lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.669671] env[61824]: DEBUG nova.compute.manager [req-30f945c6-6495-4b44-8b15-e35a3a2d15d6 req-a0189621-0f5e-4bb0-ab95-d606d69fd167 service nova] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Received event network-vif-deleted-33a84cb8-47fb-42c6-b23f-6cd67baba2ea {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.709190] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.928413] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.125461] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba2947a5-1a32-4dfc-a1b8-dc2d2f5bcb35 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.134325] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5196d5d8-ad72-49a7-9cca-b5c4f13bfcd0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.178523] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15d2745-2a2a-4706-a9ff-4b61311adea1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.190798] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c295d3-092b-4d87-be83-9ba147b51065 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.206613] env[61824]: DEBUG nova.compute.provider_tree [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.212218] env[61824]: INFO nova.scheduler.client.report [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleted allocations for instance d51370a2-86d5-4c9e-978c-f3318acfd3a7 [ 680.434097] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "refresh_cache-325a641d-2509-49e6-bfbc-0b2cb85c1ff9" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.434097] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.434097] env[61824]: DEBUG nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.434097] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 680.464150] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 680.672258] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.672702] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.719380] env[61824]: DEBUG nova.scheduler.client.report [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.726176] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "d51370a2-86d5-4c9e-978c-f3318acfd3a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.856s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.967075] env[61824]: DEBUG nova.network.neutron [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.227377] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.082s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.228259] env[61824]: ERROR nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Traceback (most recent call last): [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self.driver.spawn(context, instance, image_meta, [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] vm_ref = self.build_virtual_machine(instance, [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.228259] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] for vif in network_info: [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] return self._sync_wrapper(fn, *args, **kwargs) [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self.wait() [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self[:] = self._gt.wait() [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] return self._exit_event.wait() [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] current.throw(*self._exc) [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.228635] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] result = function(*args, **kwargs) [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] return func(*args, **kwargs) [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] raise e [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] nwinfo = self.network_api.allocate_for_instance( [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] created_port_ids = self._update_ports_for_instance( [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] with excutils.save_and_reraise_exception(): [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] self.force_reraise() [ 681.229038] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] raise self.value [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] updated_port = self._update_port( [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] _ensure_no_port_binding_failure(port) [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] raise exception.PortBindingFailed(port_id=port['id']) [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] nova.exception.PortBindingFailed: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. [ 681.229368] env[61824]: ERROR nova.compute.manager [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] [ 681.229368] env[61824]: DEBUG nova.compute.utils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 681.231329] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Build of instance b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d was re-scheduled: Binding failed for port fec9e102-4433-4cbd-8d36-a1d5dfb250c6, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 681.231831] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 681.232065] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Acquiring lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.232357] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Acquired lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.233138] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.233953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.691s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.239612] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.266485] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.396909] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.470628] env[61824]: INFO nova.compute.manager [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: 325a641d-2509-49e6-bfbc-0b2cb85c1ff9] Took 1.04 seconds to deallocate network for instance. [ 681.788114] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.901450] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Releasing lock "refresh_cache-b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.901678] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.901860] env[61824]: DEBUG nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.902064] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.920901] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.122396] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ebd795-3793-4f96-9267-797f62b8ca5d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.130450] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27dabbda-0f77-4712-8376-ac6cf47875d0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.161734] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d54201-ca17-49bb-94ba-ceacee99e8aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.168972] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf48d29-a3cf-4721-8313-f80eb74e13ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.181923] env[61824]: DEBUG nova.compute.provider_tree [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.343314] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.343691] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.423503] env[61824]: DEBUG nova.network.neutron [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.521582] env[61824]: INFO nova.scheduler.client.report [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleted allocations for instance 325a641d-2509-49e6-bfbc-0b2cb85c1ff9 [ 682.687038] env[61824]: DEBUG nova.scheduler.client.report [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.929131] env[61824]: INFO nova.compute.manager [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] [instance: b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d] Took 1.03 seconds to deallocate network for instance. [ 683.037100] env[61824]: DEBUG oslo_concurrency.lockutils [None req-115426ee-3a5e-4dd0-b864-d8c9c17ed195 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "325a641d-2509-49e6-bfbc-0b2cb85c1ff9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.130s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.192182] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.192870] env[61824]: ERROR nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Traceback (most recent call last): [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self.driver.spawn(context, instance, image_meta, [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] vm_ref = self.build_virtual_machine(instance, [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.192870] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] for vif in network_info: [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] return self._sync_wrapper(fn, *args, **kwargs) [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self.wait() [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self[:] = self._gt.wait() [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] return self._exit_event.wait() [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] current.throw(*self._exc) [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.193224] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] result = function(*args, **kwargs) [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] return func(*args, **kwargs) [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] raise e [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] nwinfo = self.network_api.allocate_for_instance( [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] created_port_ids = self._update_ports_for_instance( [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] with excutils.save_and_reraise_exception(): [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] self.force_reraise() [ 683.193603] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] raise self.value [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] updated_port = self._update_port( [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] _ensure_no_port_binding_failure(port) [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] raise exception.PortBindingFailed(port_id=port['id']) [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] nova.exception.PortBindingFailed: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. [ 683.193984] env[61824]: ERROR nova.compute.manager [instance: 3cd00254-b61b-498a-946b-ff2028230e25] [ 683.193984] env[61824]: DEBUG nova.compute.utils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.194776] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.161s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.199306] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Build of instance 3cd00254-b61b-498a-946b-ff2028230e25 was re-scheduled: Binding failed for port 1f433b02-2760-4410-9da4-72563f09ae3e, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 683.199742] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 683.199962] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Acquiring lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.200120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Acquired lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.200277] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 683.540083] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.732351] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.852570] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.955302] env[61824]: INFO nova.scheduler.client.report [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Deleted allocations for instance b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d [ 684.059316] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fad6bb9-29d8-44a4-bca7-84f272b7059f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.062386] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.067806] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-822b8317-8bb3-4c00-8b7b-50ecb4da09df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.097714] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9520ae54-aaf5-4c6b-a0e2-4d9b173f170b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.104983] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979d9680-637a-47aa-a10c-6dfae844215e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.119456] env[61824]: DEBUG nova.compute.provider_tree [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.358960] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Releasing lock "refresh_cache-3cd00254-b61b-498a-946b-ff2028230e25" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.359168] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 684.359879] env[61824]: DEBUG nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.359879] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 684.376077] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 684.389462] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "f24881d3-a647-42f9-bf82-0065f3425a5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.389715] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.412028] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "e3e19a04-a29b-4473-833c-d8486703e9ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.413440] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.462323] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e68e41f7-fc70-44cc-820e-13676dc28830 tempest-ImagesOneServerNegativeTestJSON-1773770547 tempest-ImagesOneServerNegativeTestJSON-1773770547-project-member] Lock "b4dd79b0-5dbb-4f12-9736-d8bc9f828f5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.208s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.622897] env[61824]: DEBUG nova.scheduler.client.report [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.880458] env[61824]: DEBUG nova.network.neutron [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.966881] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.128957] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.934s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.130316] env[61824]: ERROR nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Traceback (most recent call last): [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self.driver.spawn(context, instance, image_meta, [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] vm_ref = self.build_virtual_machine(instance, [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.130316] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] for vif in network_info: [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] return self._sync_wrapper(fn, *args, **kwargs) [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self.wait() [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self[:] = self._gt.wait() [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] return self._exit_event.wait() [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] current.throw(*self._exc) [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.130654] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] result = function(*args, **kwargs) [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] return func(*args, **kwargs) [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] raise e [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] nwinfo = self.network_api.allocate_for_instance( [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] created_port_ids = self._update_ports_for_instance( [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] with excutils.save_and_reraise_exception(): [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] self.force_reraise() [ 685.131113] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] raise self.value [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] updated_port = self._update_port( [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] _ensure_no_port_binding_failure(port) [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] raise exception.PortBindingFailed(port_id=port['id']) [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] nova.exception.PortBindingFailed: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. [ 685.131474] env[61824]: ERROR nova.compute.manager [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] [ 685.131474] env[61824]: DEBUG nova.compute.utils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.132752] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Build of instance e56a6950-f842-45fb-a0a0-0defe3b3c30f was re-scheduled: Binding failed for port 3d8f0991-2a49-4263-8cb1-0197df211305, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 685.133180] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 685.133406] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.133555] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.133710] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.135232] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.191s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.136920] env[61824]: INFO nova.compute.claims [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.383535] env[61824]: INFO nova.compute.manager [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] [instance: 3cd00254-b61b-498a-946b-ff2028230e25] Took 1.02 seconds to deallocate network for instance. [ 685.496537] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.657070] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 685.808759] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.312655] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-e56a6950-f842-45fb-a0a0-0defe3b3c30f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.312896] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 686.313086] env[61824]: DEBUG nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.313270] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 686.362173] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.421651] env[61824]: INFO nova.scheduler.client.report [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Deleted allocations for instance 3cd00254-b61b-498a-946b-ff2028230e25 [ 686.601992] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77eca08a-661d-4328-aa7c-d7a63f9640f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.610237] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429d81fc-442c-431a-ba66-b1d13f0815d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.640207] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98de1234-30f5-406f-8f34-0ef913f1721a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.648306] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6517fb5-d782-43db-96cc-cc9b39e83df6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.661582] env[61824]: DEBUG nova.compute.provider_tree [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.862565] env[61824]: DEBUG nova.network.neutron [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.934058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1723708-c327-4b1b-be04-808b055be64e tempest-InstanceActionsV221TestJSON-334398501 tempest-InstanceActionsV221TestJSON-334398501-project-member] Lock "3cd00254-b61b-498a-946b-ff2028230e25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.408s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.164668] env[61824]: DEBUG nova.scheduler.client.report [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.365652] env[61824]: INFO nova.compute.manager [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: e56a6950-f842-45fb-a0a0-0defe3b3c30f] Took 1.05 seconds to deallocate network for instance. [ 687.437878] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.669253] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.670847] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.672783] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.809s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.674090] env[61824]: INFO nova.compute.claims [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.963228] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.179840] env[61824]: DEBUG nova.compute.utils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.187975] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 688.188167] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 688.311704] env[61824]: DEBUG nova.policy [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e01f1c8c0e8409c9ec265aadb054769', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23c041d32dc3413281048027a3001ed2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.404772] env[61824]: INFO nova.scheduler.client.report [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted allocations for instance e56a6950-f842-45fb-a0a0-0defe3b3c30f [ 688.688457] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.917018] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9222acf9-eaa2-4284-b015-c972d3429085 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "e56a6950-f842-45fb-a0a0-0defe3b3c30f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.269s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.051932] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Successfully created port: f3e57e02-5736-4d6d-b52a-ba42f46d1689 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 689.164743] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c44691-6b55-4152-a864-2cc9df30f1cb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.178077] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a4fe9f3-cffa-42bc-a39c-9801585c3894 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.220104] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df79b9d-dafa-4acc-9869-c33713f98f0c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.229096] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9c4012-b51f-4a4a-838d-195b34774364 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.244254] env[61824]: DEBUG nova.compute.provider_tree [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.419015] env[61824]: DEBUG nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.724192] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.753027] env[61824]: DEBUG nova.scheduler.client.report [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.764766] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.764997] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.765164] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.765345] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.765485] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.766288] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.766288] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.766288] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.766472] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.766792] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.767390] env[61824]: DEBUG nova.virt.hardware [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.767952] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce68b4e-d8a1-42d8-800d-a12e30c8b20d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.781337] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16ced79-5fc3-4c74-b03e-0d3f45f75cb9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.953946] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.259437] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.259986] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.263517] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.412s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.264512] env[61824]: INFO nova.compute.claims [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.508184] env[61824]: DEBUG nova.compute.manager [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Received event network-changed-f3e57e02-5736-4d6d-b52a-ba42f46d1689 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.508342] env[61824]: DEBUG nova.compute.manager [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Refreshing instance network info cache due to event network-changed-f3e57e02-5736-4d6d-b52a-ba42f46d1689. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 690.508560] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] Acquiring lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.508702] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] Acquired lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.508862] env[61824]: DEBUG nova.network.neutron [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Refreshing network info cache for port f3e57e02-5736-4d6d-b52a-ba42f46d1689 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 690.680026] env[61824]: ERROR nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. [ 690.680026] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 690.680026] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.680026] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 690.680026] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.680026] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 690.680026] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.680026] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 690.680026] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.680026] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 690.680026] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.680026] env[61824]: ERROR nova.compute.manager raise self.value [ 690.680026] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.680026] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 690.680026] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.680026] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 690.680604] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.680604] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 690.680604] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. [ 690.680604] env[61824]: ERROR nova.compute.manager [ 690.680604] env[61824]: Traceback (most recent call last): [ 690.680604] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 690.680604] env[61824]: listener.cb(fileno) [ 690.680604] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.680604] env[61824]: result = function(*args, **kwargs) [ 690.680604] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.680604] env[61824]: return func(*args, **kwargs) [ 690.680604] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.680604] env[61824]: raise e [ 690.680604] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.680604] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 690.680604] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.680604] env[61824]: created_port_ids = self._update_ports_for_instance( [ 690.680604] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.680604] env[61824]: with excutils.save_and_reraise_exception(): [ 690.680604] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.680604] env[61824]: self.force_reraise() [ 690.680604] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.680604] env[61824]: raise self.value [ 690.680604] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.680604] env[61824]: updated_port = self._update_port( [ 690.680604] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.680604] env[61824]: _ensure_no_port_binding_failure(port) [ 690.680604] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.680604] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 690.681469] env[61824]: nova.exception.PortBindingFailed: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. [ 690.681469] env[61824]: Removing descriptor: 17 [ 690.681469] env[61824]: ERROR nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] Traceback (most recent call last): [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] yield resources [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self.driver.spawn(context, instance, image_meta, [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.681469] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] vm_ref = self.build_virtual_machine(instance, [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] for vif in network_info: [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return self._sync_wrapper(fn, *args, **kwargs) [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self.wait() [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self[:] = self._gt.wait() [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return self._exit_event.wait() [ 690.681859] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] result = hub.switch() [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return self.greenlet.switch() [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] result = function(*args, **kwargs) [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return func(*args, **kwargs) [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] raise e [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] nwinfo = self.network_api.allocate_for_instance( [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.682256] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] created_port_ids = self._update_ports_for_instance( [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] with excutils.save_and_reraise_exception(): [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self.force_reraise() [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] raise self.value [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] updated_port = self._update_port( [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] _ensure_no_port_binding_failure(port) [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.682620] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] raise exception.PortBindingFailed(port_id=port['id']) [ 690.682921] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] nova.exception.PortBindingFailed: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. [ 690.682921] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] [ 690.682921] env[61824]: INFO nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Terminating instance [ 690.683863] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.775067] env[61824]: DEBUG nova.compute.utils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.775067] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.775067] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 690.891056] env[61824]: DEBUG nova.policy [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a81282e41a294fb9bf8524dd15704cfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '92f9e969563247e1a0121e620ea5d2ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.939112] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.939112] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.027478] env[61824]: DEBUG nova.network.neutron [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.080010] env[61824]: DEBUG nova.network.neutron [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.278718] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.407410] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Successfully created port: 4ed391ee-ca92-4c4d-8cfc-04ed471a015a {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.583505] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f1e8bc9-91d7-409d-9f82-39b885ddc747 req-366eac65-2727-4504-8d82-16a8493ae660 service nova] Releasing lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.583932] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquired lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.584226] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.760414] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9ff52a-2fc9-47fa-b529-348b78e57266 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.770313] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ed793b-76c4-4e4e-b7a0-d8a4c30f60de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.813169] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c64ff79-615c-4b4d-8103-51f88647f05e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.823614] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9da3c73-4d46-4671-a970-e135866ffe9a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.839325] env[61824]: DEBUG nova.compute.provider_tree [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.111814] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.159096] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.314537] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "8522fe80-4d6d-4474-ae63-00daa0dc7349" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.314779] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.317259] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.342385] env[61824]: DEBUG nova.scheduler.client.report [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.360361] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:53:46Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='b6536f04-d05f-4270-a910-6ffc3abac317',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1106474739',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.360728] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.360801] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.360914] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.361078] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.361229] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.361431] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.361588] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.361764] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.362114] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.362344] env[61824]: DEBUG nova.virt.hardware [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.363951] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae95326c-4034-4860-965d-a28641bcc08e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.373406] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95501ebd-778d-45b4-903b-b52df12e469a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.539090] env[61824]: DEBUG nova.compute.manager [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Received event network-vif-deleted-f3e57e02-5736-4d6d-b52a-ba42f46d1689 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.539302] env[61824]: DEBUG nova.compute.manager [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Received event network-changed-4ed391ee-ca92-4c4d-8cfc-04ed471a015a {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.539463] env[61824]: DEBUG nova.compute.manager [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Refreshing instance network info cache due to event network-changed-4ed391ee-ca92-4c4d-8cfc-04ed471a015a. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.539939] env[61824]: DEBUG oslo_concurrency.lockutils [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] Acquiring lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.540061] env[61824]: DEBUG oslo_concurrency.lockutils [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] Acquired lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.540224] env[61824]: DEBUG nova.network.neutron [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Refreshing network info cache for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 692.662620] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Releasing lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.663638] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 692.663859] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 692.664222] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bba9827-1933-4566-9699-f1d69dedc943 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.675873] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec4d6fd-a00c-4fb5-9dfb-aac9fb71427f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.702016] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 717eca7a-0287-4424-a302-5737f92724a6 could not be found. [ 692.702279] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 692.702460] env[61824]: INFO nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 692.702714] env[61824]: DEBUG oslo.service.loopingcall [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.702926] env[61824]: DEBUG nova.compute.manager [-] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 692.703026] env[61824]: DEBUG nova.network.neutron [-] [instance: 717eca7a-0287-4424-a302-5737f92724a6] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 692.719678] env[61824]: DEBUG nova.network.neutron [-] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.722136] env[61824]: ERROR nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. [ 692.722136] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.722136] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.722136] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.722136] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.722136] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.722136] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.722136] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.722136] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.722136] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 692.722136] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.722136] env[61824]: ERROR nova.compute.manager raise self.value [ 692.722136] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.722136] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.722136] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.722136] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.722590] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.722590] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.722590] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. [ 692.722590] env[61824]: ERROR nova.compute.manager [ 692.722590] env[61824]: Traceback (most recent call last): [ 692.722590] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.722590] env[61824]: listener.cb(fileno) [ 692.722590] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.722590] env[61824]: result = function(*args, **kwargs) [ 692.722590] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.722590] env[61824]: return func(*args, **kwargs) [ 692.722590] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.722590] env[61824]: raise e [ 692.722590] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.722590] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 692.722590] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.722590] env[61824]: created_port_ids = self._update_ports_for_instance( [ 692.722590] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.722590] env[61824]: with excutils.save_and_reraise_exception(): [ 692.722590] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.722590] env[61824]: self.force_reraise() [ 692.722590] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.722590] env[61824]: raise self.value [ 692.722590] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.722590] env[61824]: updated_port = self._update_port( [ 692.722590] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.722590] env[61824]: _ensure_no_port_binding_failure(port) [ 692.722590] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.722590] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.723334] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. [ 692.723334] env[61824]: Removing descriptor: 17 [ 692.723334] env[61824]: ERROR nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Traceback (most recent call last): [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] yield resources [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self.driver.spawn(context, instance, image_meta, [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.723334] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] vm_ref = self.build_virtual_machine(instance, [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] for vif in network_info: [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return self._sync_wrapper(fn, *args, **kwargs) [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self.wait() [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self[:] = self._gt.wait() [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return self._exit_event.wait() [ 692.723639] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] result = hub.switch() [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return self.greenlet.switch() [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] result = function(*args, **kwargs) [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return func(*args, **kwargs) [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] raise e [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] nwinfo = self.network_api.allocate_for_instance( [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.723974] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] created_port_ids = self._update_ports_for_instance( [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] with excutils.save_and_reraise_exception(): [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self.force_reraise() [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] raise self.value [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] updated_port = self._update_port( [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] _ensure_no_port_binding_failure(port) [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.724363] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] raise exception.PortBindingFailed(port_id=port['id']) [ 692.724688] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] nova.exception.PortBindingFailed: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. [ 692.724688] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] [ 692.724688] env[61824]: INFO nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Terminating instance [ 692.729283] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.851773] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.854067] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 692.857244] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.098s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.068885] env[61824]: DEBUG nova.network.neutron [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.146055] env[61824]: DEBUG nova.network.neutron [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.225513] env[61824]: DEBUG nova.network.neutron [-] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.367022] env[61824]: DEBUG nova.compute.utils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.368068] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 693.368568] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.423476] env[61824]: DEBUG nova.policy [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e01f1c8c0e8409c9ec265aadb054769', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '23c041d32dc3413281048027a3001ed2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.650463] env[61824]: DEBUG oslo_concurrency.lockutils [req-15f3c78a-a374-4167-bf96-bf1f0ed4b3cf req-ac750e75-7f38-44ab-9e3c-cfd05a7d74ff service nova] Releasing lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.650738] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquired lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.650920] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 693.681083] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075ff6a0-5c85-4c82-8c43-e38d08794254 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.689801] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79a5d71-ab48-4349-b64f-c43f96dfda48 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.722125] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b801bcb-666a-4d41-ae19-31ba9de901e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.729422] env[61824]: INFO nova.compute.manager [-] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Took 1.03 seconds to deallocate network for instance. [ 693.733349] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d82024-9f8a-40ab-bd8c-e4e42a48acaa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.739196] env[61824]: DEBUG nova.compute.claims [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 693.739520] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.750829] env[61824]: DEBUG nova.compute.provider_tree [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.756018] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Successfully created port: fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.871642] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 694.171052] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.223546] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.255936] env[61824]: DEBUG nova.scheduler.client.report [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.726277] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Releasing lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.726761] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.726948] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.727273] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b958608f-6e24-44ba-b056-b11604e0c35c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.740637] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0063435e-25ee-47b0-954d-1e6313a133aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.760439] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.761184] env[61824]: ERROR nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Traceback (most recent call last): [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self.driver.spawn(context, instance, image_meta, [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] vm_ref = self.build_virtual_machine(instance, [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.761184] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] for vif in network_info: [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] return self._sync_wrapper(fn, *args, **kwargs) [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self.wait() [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self[:] = self._gt.wait() [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] return self._exit_event.wait() [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] current.throw(*self._exc) [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.761545] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] result = function(*args, **kwargs) [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] return func(*args, **kwargs) [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] raise e [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] nwinfo = self.network_api.allocate_for_instance( [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] created_port_ids = self._update_ports_for_instance( [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] with excutils.save_and_reraise_exception(): [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] self.force_reraise() [ 694.761928] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] raise self.value [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] updated_port = self._update_port( [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] _ensure_no_port_binding_failure(port) [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] raise exception.PortBindingFailed(port_id=port['id']) [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] nova.exception.PortBindingFailed: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. [ 694.762358] env[61824]: ERROR nova.compute.manager [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] [ 694.762358] env[61824]: DEBUG nova.compute.utils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.766967] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.238s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.769911] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8aeeb8e9-e1ab-463b-bc16-ca68571882aa could not be found. [ 694.770121] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.770300] env[61824]: INFO nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Took 0.04 seconds to destroy the instance on the hypervisor. [ 694.770537] env[61824]: DEBUG oslo.service.loopingcall [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.771128] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Build of instance 9ce52867-d547-401e-993d-0053eb9da4c4 was re-scheduled: Binding failed for port 35ea00d6-de28-4d0c-9fcb-92fca8a21068, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 694.771528] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 694.771937] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Acquiring lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.771937] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Acquired lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.772062] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 694.772886] env[61824]: DEBUG nova.compute.manager [-] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.773037] env[61824]: DEBUG nova.network.neutron [-] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 694.793819] env[61824]: DEBUG nova.network.neutron [-] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.798920] env[61824]: DEBUG nova.compute.manager [req-a1d84313-5b5e-4d7e-9d71-f5bf36d8eaf9 req-0f9b45b6-cd3e-4920-a0f8-d07fe0e7ae72 service nova] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Received event network-vif-deleted-4ed391ee-ca92-4c4d-8cfc-04ed471a015a {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.881275] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 694.911147] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 694.911147] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 694.911147] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.911317] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 694.911317] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.911317] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 694.911317] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 694.911317] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 694.914083] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 694.914249] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 694.914426] env[61824]: DEBUG nova.virt.hardware [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 694.915302] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90b28b0-6325-4e94-88cb-7c921c589f63 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.926500] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97c645d-3a5d-4343-ab48-efd3dbb373ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.200455] env[61824]: ERROR nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. [ 695.200455] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.200455] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.200455] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.200455] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.200455] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.200455] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.200455] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.200455] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.200455] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 695.200455] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.200455] env[61824]: ERROR nova.compute.manager raise self.value [ 695.200455] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.200455] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.200455] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.200455] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.200973] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.200973] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.200973] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. [ 695.200973] env[61824]: ERROR nova.compute.manager [ 695.200973] env[61824]: Traceback (most recent call last): [ 695.200973] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.200973] env[61824]: listener.cb(fileno) [ 695.200973] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.200973] env[61824]: result = function(*args, **kwargs) [ 695.200973] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.200973] env[61824]: return func(*args, **kwargs) [ 695.200973] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.200973] env[61824]: raise e [ 695.200973] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.200973] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 695.200973] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.200973] env[61824]: created_port_ids = self._update_ports_for_instance( [ 695.200973] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.200973] env[61824]: with excutils.save_and_reraise_exception(): [ 695.200973] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.200973] env[61824]: self.force_reraise() [ 695.200973] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.200973] env[61824]: raise self.value [ 695.200973] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.200973] env[61824]: updated_port = self._update_port( [ 695.200973] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.200973] env[61824]: _ensure_no_port_binding_failure(port) [ 695.200973] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.200973] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.201813] env[61824]: nova.exception.PortBindingFailed: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. [ 695.201813] env[61824]: Removing descriptor: 17 [ 695.201813] env[61824]: ERROR nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Traceback (most recent call last): [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] yield resources [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self.driver.spawn(context, instance, image_meta, [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.201813] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] vm_ref = self.build_virtual_machine(instance, [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] for vif in network_info: [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return self._sync_wrapper(fn, *args, **kwargs) [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self.wait() [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self[:] = self._gt.wait() [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return self._exit_event.wait() [ 695.202204] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] result = hub.switch() [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return self.greenlet.switch() [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] result = function(*args, **kwargs) [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return func(*args, **kwargs) [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] raise e [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] nwinfo = self.network_api.allocate_for_instance( [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 695.202582] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] created_port_ids = self._update_ports_for_instance( [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] with excutils.save_and_reraise_exception(): [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self.force_reraise() [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] raise self.value [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] updated_port = self._update_port( [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] _ensure_no_port_binding_failure(port) [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.202976] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] raise exception.PortBindingFailed(port_id=port['id']) [ 695.203347] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] nova.exception.PortBindingFailed: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. [ 695.203347] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] [ 695.203347] env[61824]: INFO nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Terminating instance [ 695.203670] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.203783] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquired lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.203944] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.291641] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.296063] env[61824]: DEBUG nova.network.neutron [-] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.370320] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.502101] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Acquiring lock "6235a9b9-19c0-4917-938a-bca82a2ca05e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.502297] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.617575] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47139b00-055a-494d-a8d3-2883631d54ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.626113] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3d0652-f444-489e-ac3b-41e5a2e31c9d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.657999] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43219bc-7857-4705-be4e-67e0adf56620 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.666054] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de84b422-336a-4f6b-9b85-2b6efb7fd456 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.679882] env[61824]: DEBUG nova.compute.provider_tree [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.731384] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.799105] env[61824]: INFO nova.compute.manager [-] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Took 1.03 seconds to deallocate network for instance. [ 695.801559] env[61824]: DEBUG nova.compute.claims [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.801800] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.808901] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.872524] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Releasing lock "refresh_cache-9ce52867-d547-401e-993d-0053eb9da4c4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.872749] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 695.872928] env[61824]: DEBUG nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.873117] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 695.887669] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.183443] env[61824]: DEBUG nova.scheduler.client.report [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.311820] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Releasing lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.312419] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.312685] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 696.313086] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fd7b4f5-d33c-4c30-802f-5bce8f4700a7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.324527] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155e5281-1124-48fd-9711-93bdd8ad47ad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.347215] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e could not be found. [ 696.347616] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 696.347616] env[61824]: INFO nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 696.347855] env[61824]: DEBUG oslo.service.loopingcall [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.348066] env[61824]: DEBUG nova.compute.manager [-] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 696.348158] env[61824]: DEBUG nova.network.neutron [-] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.361830] env[61824]: DEBUG nova.network.neutron [-] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.390086] env[61824]: DEBUG nova.network.neutron [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.688634] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.921s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.689454] env[61824]: ERROR nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Traceback (most recent call last): [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self.driver.spawn(context, instance, image_meta, [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] vm_ref = self.build_virtual_machine(instance, [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.689454] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] for vif in network_info: [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return self._sync_wrapper(fn, *args, **kwargs) [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self.wait() [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self[:] = self._gt.wait() [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return self._exit_event.wait() [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] result = hub.switch() [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 696.689833] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return self.greenlet.switch() [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] result = function(*args, **kwargs) [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] return func(*args, **kwargs) [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] raise e [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] nwinfo = self.network_api.allocate_for_instance( [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] created_port_ids = self._update_ports_for_instance( [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] with excutils.save_and_reraise_exception(): [ 696.690224] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] self.force_reraise() [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] raise self.value [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] updated_port = self._update_port( [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] _ensure_no_port_binding_failure(port) [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] raise exception.PortBindingFailed(port_id=port['id']) [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] nova.exception.PortBindingFailed: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. [ 696.690610] env[61824]: ERROR nova.compute.manager [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] [ 696.690990] env[61824]: DEBUG nova.compute.utils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.691444] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.904s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.693026] env[61824]: INFO nova.compute.claims [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.695734] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Build of instance 7a93d1c2-20d6-438e-bc42-1f9aa0219f91 was re-scheduled: Binding failed for port 33a84cb8-47fb-42c6-b23f-6cd67baba2ea, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 696.696184] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 696.696406] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquiring lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.696593] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Acquired lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.696758] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 696.840133] env[61824]: DEBUG nova.compute.manager [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Received event network-changed-fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.840328] env[61824]: DEBUG nova.compute.manager [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Refreshing instance network info cache due to event network-changed-fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 696.840542] env[61824]: DEBUG oslo_concurrency.lockutils [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] Acquiring lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.840682] env[61824]: DEBUG oslo_concurrency.lockutils [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] Acquired lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.840834] env[61824]: DEBUG nova.network.neutron [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Refreshing network info cache for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 696.864961] env[61824]: DEBUG nova.network.neutron [-] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.892572] env[61824]: INFO nova.compute.manager [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] [instance: 9ce52867-d547-401e-993d-0053eb9da4c4] Took 1.02 seconds to deallocate network for instance. [ 697.222487] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.323879] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.362122] env[61824]: DEBUG nova.network.neutron [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.367441] env[61824]: INFO nova.compute.manager [-] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Took 1.02 seconds to deallocate network for instance. [ 697.369758] env[61824]: DEBUG nova.compute.claims [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 697.369930] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.435503] env[61824]: DEBUG nova.network.neutron [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.826286] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Releasing lock "refresh_cache-7a93d1c2-20d6-438e-bc42-1f9aa0219f91" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.826583] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 697.826807] env[61824]: DEBUG nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.826987] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.848784] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.924141] env[61824]: INFO nova.scheduler.client.report [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Deleted allocations for instance 9ce52867-d547-401e-993d-0053eb9da4c4 [ 697.938191] env[61824]: DEBUG oslo_concurrency.lockutils [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] Releasing lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.938191] env[61824]: DEBUG nova.compute.manager [req-d3990671-7e7b-46a0-91c5-070602cb5004 req-f1978de3-fe40-45a0-ab3c-537209432267 service nova] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Received event network-vif-deleted-fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 698.047951] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43aa33b9-5f6f-487b-8e38-46c4081dad6c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.053500] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcdb9b1f-3257-4e3c-9420-215ad47c2a39 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.084423] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b8f6d9-d33e-4bef-b3da-7f81839071ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.092603] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104568ca-0be5-4449-8beb-c1ad3c131105 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.106076] env[61824]: DEBUG nova.compute.provider_tree [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.352401] env[61824]: DEBUG nova.network.neutron [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.435045] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e272b1d7-b455-45ac-8d37-36f922970fc6 tempest-ServersV294TestFqdnHostnames-599466120 tempest-ServersV294TestFqdnHostnames-599466120-project-member] Lock "9ce52867-d547-401e-993d-0053eb9da4c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.188s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.608700] env[61824]: DEBUG nova.scheduler.client.report [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.854266] env[61824]: INFO nova.compute.manager [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] [instance: 7a93d1c2-20d6-438e-bc42-1f9aa0219f91] Took 1.03 seconds to deallocate network for instance. [ 698.937835] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 699.113740] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.114838] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 699.117159] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.055s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.118686] env[61824]: INFO nova.compute.claims [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.465142] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.625445] env[61824]: DEBUG nova.compute.utils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.629043] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 699.629217] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 699.681326] env[61824]: DEBUG nova.policy [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a975aca3353469a948c52b035e17fcc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '502b41859ba843838d2188c82f70ae08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.892578] env[61824]: INFO nova.scheduler.client.report [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Deleted allocations for instance 7a93d1c2-20d6-438e-bc42-1f9aa0219f91 [ 700.023733] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Successfully created port: 1c4e8983-546f-4c22-98cd-5f12b4c8322d {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 700.130177] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 700.400553] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ac06b859-9cfd-4a51-a958-f22bc1767b1e tempest-SecurityGroupsTestJSON-1378887067 tempest-SecurityGroupsTestJSON-1378887067-project-member] Lock "7a93d1c2-20d6-438e-bc42-1f9aa0219f91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.496s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.477401] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f4f5c2b-ae7a-4bc0-80e5-bbf287cc17f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.485132] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885aa65e-c167-4db1-9893-583bf1b0ae63 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.520120] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155e5d66-9ee2-4bc8-8d4e-c3987cda649a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.528780] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a30ad61-5e54-4228-b8d8-8c42056f338a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.544247] env[61824]: DEBUG nova.compute.provider_tree [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.903586] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 701.049690] env[61824]: DEBUG nova.scheduler.client.report [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 701.143364] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 701.177338] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 701.177596] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 701.177776] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 701.177961] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 701.181258] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 701.181446] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 701.181657] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 701.181820] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 701.181986] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 701.182161] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 701.182334] env[61824]: DEBUG nova.virt.hardware [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 701.183203] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f111fd28-beb9-4fc1-8e35-b40e69c21281 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.193438] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f14b14-5fe8-4e8b-b779-f169d5723f50 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.433430] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.560018] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.560018] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.561778] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.065s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.563306] env[61824]: INFO nova.compute.claims [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.697306] env[61824]: DEBUG nova.compute.manager [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Received event network-changed-1c4e8983-546f-4c22-98cd-5f12b4c8322d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.697736] env[61824]: DEBUG nova.compute.manager [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Refreshing instance network info cache due to event network-changed-1c4e8983-546f-4c22-98cd-5f12b4c8322d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 701.698103] env[61824]: DEBUG oslo_concurrency.lockutils [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] Acquiring lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.699923] env[61824]: DEBUG oslo_concurrency.lockutils [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] Acquired lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.699923] env[61824]: DEBUG nova.network.neutron [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Refreshing network info cache for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.914127] env[61824]: ERROR nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. [ 701.914127] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 701.914127] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.914127] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 701.914127] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.914127] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 701.914127] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.914127] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 701.914127] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.914127] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 701.914127] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.914127] env[61824]: ERROR nova.compute.manager raise self.value [ 701.914127] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.914127] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 701.914127] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.914127] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 701.914647] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.914647] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 701.914647] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. [ 701.914647] env[61824]: ERROR nova.compute.manager [ 701.914647] env[61824]: Traceback (most recent call last): [ 701.914647] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 701.914647] env[61824]: listener.cb(fileno) [ 701.914647] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.914647] env[61824]: result = function(*args, **kwargs) [ 701.914647] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.914647] env[61824]: return func(*args, **kwargs) [ 701.914647] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.914647] env[61824]: raise e [ 701.914647] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.914647] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 701.914647] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.914647] env[61824]: created_port_ids = self._update_ports_for_instance( [ 701.914647] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.914647] env[61824]: with excutils.save_and_reraise_exception(): [ 701.914647] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.914647] env[61824]: self.force_reraise() [ 701.914647] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.914647] env[61824]: raise self.value [ 701.914647] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.914647] env[61824]: updated_port = self._update_port( [ 701.914647] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.914647] env[61824]: _ensure_no_port_binding_failure(port) [ 701.914647] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.914647] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 701.915463] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. [ 701.915463] env[61824]: Removing descriptor: 17 [ 701.915463] env[61824]: ERROR nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Traceback (most recent call last): [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] yield resources [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self.driver.spawn(context, instance, image_meta, [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.915463] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] vm_ref = self.build_virtual_machine(instance, [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] for vif in network_info: [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return self._sync_wrapper(fn, *args, **kwargs) [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self.wait() [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self[:] = self._gt.wait() [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return self._exit_event.wait() [ 701.915821] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] result = hub.switch() [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return self.greenlet.switch() [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] result = function(*args, **kwargs) [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return func(*args, **kwargs) [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] raise e [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] nwinfo = self.network_api.allocate_for_instance( [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 701.916727] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] created_port_ids = self._update_ports_for_instance( [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] with excutils.save_and_reraise_exception(): [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self.force_reraise() [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] raise self.value [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] updated_port = self._update_port( [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] _ensure_no_port_binding_failure(port) [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.917159] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] raise exception.PortBindingFailed(port_id=port['id']) [ 701.918026] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] nova.exception.PortBindingFailed: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. [ 701.918026] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] [ 701.918026] env[61824]: INFO nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Terminating instance [ 701.919149] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.069097] env[61824]: DEBUG nova.compute.utils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.073850] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 702.073935] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 702.126815] env[61824]: DEBUG nova.policy [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0941c0e9a43b44a396ac57e7b8eaa45d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b030eebd56bb416fb6507640c875f416', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.219270] env[61824]: DEBUG nova.network.neutron [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.305522] env[61824]: DEBUG nova.network.neutron [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.577926] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.627806] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Successfully created port: caf0c180-8fb5-411d-b711-0075a7403c83 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.807894] env[61824]: DEBUG oslo_concurrency.lockutils [req-a22e64a4-bc19-4f54-a2b5-d47509d20054 req-0ebad191-9b1e-4771-90cd-32a413af1333 service nova] Releasing lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.808336] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.808525] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 702.959647] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464ee0e5-e382-41a8-8a6b-ae4d7a80e2b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.970554] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0094766-bf29-48d7-9cd4-27db261fbaad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.004142] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb5cf10-3d08-49c2-90e9-a370d48cf4cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.012672] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a284bf55-a53d-4749-bbba-35028ed7bcac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.026858] env[61824]: DEBUG nova.compute.provider_tree [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.329501] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.446874] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.533439] env[61824]: DEBUG nova.scheduler.client.report [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.588479] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.623468] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.623742] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.623906] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.624257] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.624464] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.624641] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.624878] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.625094] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.625303] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.625495] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.625696] env[61824]: DEBUG nova.virt.hardware [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.627008] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dcfbe5d-2e8a-4302-afe8-eee54bcf596a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.635622] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7069c88b-1c06-4295-9907-eff0d613ec07 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.733517] env[61824]: DEBUG nova.compute.manager [req-db5e28d6-39da-498d-b8ec-398f2116a135 req-8a45c7c7-7b34-4fd1-b832-57b3cadeb184 service nova] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Received event network-vif-deleted-1c4e8983-546f-4c22-98cd-5f12b4c8322d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 703.952673] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.953407] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.954175] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.954319] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-485cc2a8-25dc-41e9-ae41-7c9b9be8833c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.967704] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d176b1-7810-4363-a65d-2185f9ed43a5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.999078] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 037dd0ba-0d50-4c46-8947-7df3ca19e1c1 could not be found. [ 703.999375] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.000128] env[61824]: INFO nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Took 0.05 seconds to destroy the instance on the hypervisor. [ 704.000450] env[61824]: DEBUG oslo.service.loopingcall [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.000922] env[61824]: DEBUG nova.compute.manager [-] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.001644] env[61824]: DEBUG nova.network.neutron [-] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.033662] env[61824]: DEBUG nova.network.neutron [-] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.038274] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.038723] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 704.041310] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.078s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.042848] env[61824]: INFO nova.compute.claims [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.057697] env[61824]: ERROR nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. [ 704.057697] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 704.057697] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.057697] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 704.057697] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 704.057697] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 704.057697] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 704.057697] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 704.057697] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.057697] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 704.057697] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.057697] env[61824]: ERROR nova.compute.manager raise self.value [ 704.057697] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 704.057697] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 704.057697] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.057697] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 704.058146] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.058146] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 704.058146] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. [ 704.058146] env[61824]: ERROR nova.compute.manager [ 704.058146] env[61824]: Traceback (most recent call last): [ 704.058146] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 704.058146] env[61824]: listener.cb(fileno) [ 704.058146] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.058146] env[61824]: result = function(*args, **kwargs) [ 704.058146] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.058146] env[61824]: return func(*args, **kwargs) [ 704.058146] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.058146] env[61824]: raise e [ 704.058146] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.058146] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 704.058146] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 704.058146] env[61824]: created_port_ids = self._update_ports_for_instance( [ 704.058146] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 704.058146] env[61824]: with excutils.save_and_reraise_exception(): [ 704.058146] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.058146] env[61824]: self.force_reraise() [ 704.058146] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.058146] env[61824]: raise self.value [ 704.058146] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 704.058146] env[61824]: updated_port = self._update_port( [ 704.058146] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.058146] env[61824]: _ensure_no_port_binding_failure(port) [ 704.058146] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.058146] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 704.058872] env[61824]: nova.exception.PortBindingFailed: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. [ 704.058872] env[61824]: Removing descriptor: 17 [ 704.058872] env[61824]: ERROR nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Traceback (most recent call last): [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] yield resources [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self.driver.spawn(context, instance, image_meta, [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.058872] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] vm_ref = self.build_virtual_machine(instance, [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] for vif in network_info: [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return self._sync_wrapper(fn, *args, **kwargs) [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self.wait() [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self[:] = self._gt.wait() [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return self._exit_event.wait() [ 704.059225] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] result = hub.switch() [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return self.greenlet.switch() [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] result = function(*args, **kwargs) [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return func(*args, **kwargs) [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] raise e [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] nwinfo = self.network_api.allocate_for_instance( [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 704.059581] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] created_port_ids = self._update_ports_for_instance( [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] with excutils.save_and_reraise_exception(): [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self.force_reraise() [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] raise self.value [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] updated_port = self._update_port( [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] _ensure_no_port_binding_failure(port) [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.059960] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] raise exception.PortBindingFailed(port_id=port['id']) [ 704.060288] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] nova.exception.PortBindingFailed: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. [ 704.060288] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] [ 704.060288] env[61824]: INFO nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Terminating instance [ 704.061359] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Acquiring lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.061580] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Acquired lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.061780] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.538130] env[61824]: DEBUG nova.network.neutron [-] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.548087] env[61824]: DEBUG nova.compute.utils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.551525] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.551713] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.582133] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.644176] env[61824]: DEBUG nova.policy [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b131826ac644a639a28c4db774bee92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ad43bb71692464ea67ffe9519858f51', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.660252] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.951162] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Successfully created port: 5cb96a47-c415-423a-845e-fc3e1f820a35 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.041813] env[61824]: INFO nova.compute.manager [-] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Took 1.04 seconds to deallocate network for instance. [ 705.044173] env[61824]: DEBUG nova.compute.claims [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 705.044440] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.052075] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 705.162757] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Releasing lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.163145] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 705.163342] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.163702] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa164c6d-7229-4150-afbf-eb6a8ec11344 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.183933] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548cb652-2795-4022-8a2b-2864b4e92e37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.216080] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6e2190ac-d287-4810-9d43-740e62ae7b56 could not be found. [ 705.217026] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 705.221151] env[61824]: INFO nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Took 0.05 seconds to destroy the instance on the hypervisor. [ 705.221151] env[61824]: DEBUG oslo.service.loopingcall [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.221151] env[61824]: DEBUG nova.compute.manager [-] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.221151] env[61824]: DEBUG nova.network.neutron [-] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 705.240413] env[61824]: DEBUG nova.network.neutron [-] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.469974] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13f2287-4454-4faa-b371-9d5a6c4e84f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.477122] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c4e4e1-bdd3-4397-ae26-eac237118dfc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.512018] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f69a17a-5ceb-468b-868c-71118b4cd903 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.523455] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6c2d43-c9d7-476a-a7cf-c8e83b45966d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.539503] env[61824]: DEBUG nova.compute.provider_tree [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.744279] env[61824]: DEBUG nova.network.neutron [-] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.854068] env[61824]: DEBUG nova.compute.manager [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Received event network-changed-caf0c180-8fb5-411d-b711-0075a7403c83 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.854171] env[61824]: DEBUG nova.compute.manager [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Refreshing instance network info cache due to event network-changed-caf0c180-8fb5-411d-b711-0075a7403c83. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 705.854349] env[61824]: DEBUG oslo_concurrency.lockutils [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] Acquiring lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.854495] env[61824]: DEBUG oslo_concurrency.lockutils [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] Acquired lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.854686] env[61824]: DEBUG nova.network.neutron [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Refreshing network info cache for port caf0c180-8fb5-411d-b711-0075a7403c83 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 706.042699] env[61824]: DEBUG nova.scheduler.client.report [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.064030] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 706.094989] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 706.095266] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 706.095421] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.095601] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 706.095749] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.095891] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 706.096213] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 706.096428] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 706.096690] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 706.096906] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 706.097098] env[61824]: DEBUG nova.virt.hardware [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 706.098214] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45729eae-0272-42a2-a5ec-a3fa65b06e43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.107834] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23a3919-b36a-4a41-b361-c2c517772770 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.247162] env[61824]: INFO nova.compute.manager [-] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Took 1.03 seconds to deallocate network for instance. [ 706.249872] env[61824]: DEBUG nova.compute.claims [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 706.249872] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.429536] env[61824]: ERROR nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. [ 706.429536] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 706.429536] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.429536] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 706.429536] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.429536] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 706.429536] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.429536] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 706.429536] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.429536] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 706.429536] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.429536] env[61824]: ERROR nova.compute.manager raise self.value [ 706.429536] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.429536] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 706.429536] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.429536] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 706.430045] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.430045] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 706.430045] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. [ 706.430045] env[61824]: ERROR nova.compute.manager [ 706.430045] env[61824]: Traceback (most recent call last): [ 706.430045] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 706.430045] env[61824]: listener.cb(fileno) [ 706.430045] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.430045] env[61824]: result = function(*args, **kwargs) [ 706.430045] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.430045] env[61824]: return func(*args, **kwargs) [ 706.430045] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.430045] env[61824]: raise e [ 706.430045] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.430045] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 706.430045] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.430045] env[61824]: created_port_ids = self._update_ports_for_instance( [ 706.430045] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.430045] env[61824]: with excutils.save_and_reraise_exception(): [ 706.430045] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.430045] env[61824]: self.force_reraise() [ 706.430045] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.430045] env[61824]: raise self.value [ 706.430045] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.430045] env[61824]: updated_port = self._update_port( [ 706.430045] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.430045] env[61824]: _ensure_no_port_binding_failure(port) [ 706.430045] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.430045] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 706.430899] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. [ 706.430899] env[61824]: Removing descriptor: 17 [ 706.430899] env[61824]: ERROR nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Traceback (most recent call last): [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] yield resources [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self.driver.spawn(context, instance, image_meta, [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 706.430899] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] vm_ref = self.build_virtual_machine(instance, [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] vif_infos = vmwarevif.get_vif_info(self._session, [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] for vif in network_info: [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return self._sync_wrapper(fn, *args, **kwargs) [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self.wait() [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self[:] = self._gt.wait() [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return self._exit_event.wait() [ 706.431276] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] result = hub.switch() [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return self.greenlet.switch() [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] result = function(*args, **kwargs) [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return func(*args, **kwargs) [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] raise e [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] nwinfo = self.network_api.allocate_for_instance( [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 706.431721] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] created_port_ids = self._update_ports_for_instance( [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] with excutils.save_and_reraise_exception(): [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self.force_reraise() [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] raise self.value [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] updated_port = self._update_port( [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] _ensure_no_port_binding_failure(port) [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 706.432120] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] raise exception.PortBindingFailed(port_id=port['id']) [ 706.432549] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] nova.exception.PortBindingFailed: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. [ 706.432549] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] [ 706.432549] env[61824]: INFO nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Terminating instance [ 706.433052] env[61824]: DEBUG nova.network.neutron [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.435765] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.436082] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquired lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.436287] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 706.547611] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.548191] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.550767] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.597s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.554093] env[61824]: INFO nova.compute.claims [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.572804] env[61824]: DEBUG nova.network.neutron [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.960188] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.058765] env[61824]: DEBUG nova.compute.utils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.066464] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.066739] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 707.074849] env[61824]: DEBUG oslo_concurrency.lockutils [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] Releasing lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.075279] env[61824]: DEBUG nova.compute.manager [req-73bca622-08cc-41c1-a0ac-d51e3ad300e3 req-96f2491c-a69a-4cb3-a2fc-b70253ad3173 service nova] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Received event network-vif-deleted-caf0c180-8fb5-411d-b711-0075a7403c83 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.090882] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.138639] env[61824]: DEBUG nova.policy [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b131826ac644a639a28c4db774bee92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ad43bb71692464ea67ffe9519858f51', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.517677] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Successfully created port: 5c15fa5e-854c-4d05-922b-d032140c564c {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.567275] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.598692] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Releasing lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.599144] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.599346] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.599658] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0de49854-5cce-4a26-bfbb-7b17af2189aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.621054] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41eb950f-d5ed-43a0-914a-cacb041ca1cb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.657035] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc662f50-f111-4d26-b017-3ab719004ac2 could not be found. [ 707.657035] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.657035] env[61824]: INFO nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Took 0.06 seconds to destroy the instance on the hypervisor. [ 707.657544] env[61824]: DEBUG oslo.service.loopingcall [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.657658] env[61824]: DEBUG nova.compute.manager [-] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.657767] env[61824]: DEBUG nova.network.neutron [-] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 707.673769] env[61824]: DEBUG nova.network.neutron [-] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.964012] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7a82b6-9a95-4418-b1cf-435289b4fbe3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.971984] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bed73d-c353-44cc-9792-227d645f60d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.004357] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac20481-3fea-422f-8eac-5b6457da076f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.008066] env[61824]: DEBUG nova.compute.manager [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Received event network-changed-5cb96a47-c415-423a-845e-fc3e1f820a35 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 708.008223] env[61824]: DEBUG nova.compute.manager [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Refreshing instance network info cache due to event network-changed-5cb96a47-c415-423a-845e-fc3e1f820a35. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 708.008402] env[61824]: DEBUG oslo_concurrency.lockutils [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] Acquiring lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.008540] env[61824]: DEBUG oslo_concurrency.lockutils [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] Acquired lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.008695] env[61824]: DEBUG nova.network.neutron [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Refreshing network info cache for port 5cb96a47-c415-423a-845e-fc3e1f820a35 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 708.015801] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d227004-acc6-4284-8540-cbaa11f91286 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.030340] env[61824]: DEBUG nova.compute.provider_tree [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.177191] env[61824]: DEBUG nova.network.neutron [-] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.526938] env[61824]: DEBUG nova.network.neutron [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.536359] env[61824]: DEBUG nova.scheduler.client.report [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.576968] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.605959] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.606222] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.606377] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.606552] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.606726] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.606878] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.607092] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.607253] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.607428] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.607639] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.607783] env[61824]: DEBUG nova.virt.hardware [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.608647] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80595183-d083-4929-9279-c08824d45525 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.619514] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32e375e-5140-4d0a-8db1-d29eb1c62993 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.624067] env[61824]: DEBUG nova.network.neutron [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.681515] env[61824]: INFO nova.compute.manager [-] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Took 1.02 seconds to deallocate network for instance. [ 708.682722] env[61824]: DEBUG nova.compute.claims [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 708.683159] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.779880] env[61824]: ERROR nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. [ 708.779880] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.779880] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.779880] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.779880] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.779880] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.779880] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.779880] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.779880] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.779880] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 708.779880] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.779880] env[61824]: ERROR nova.compute.manager raise self.value [ 708.779880] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.779880] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.779880] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.779880] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.781172] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.781172] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.781172] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. [ 708.781172] env[61824]: ERROR nova.compute.manager [ 708.781172] env[61824]: Traceback (most recent call last): [ 708.781172] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.781172] env[61824]: listener.cb(fileno) [ 708.781172] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.781172] env[61824]: result = function(*args, **kwargs) [ 708.781172] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.781172] env[61824]: return func(*args, **kwargs) [ 708.781172] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.781172] env[61824]: raise e [ 708.781172] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.781172] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 708.781172] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.781172] env[61824]: created_port_ids = self._update_ports_for_instance( [ 708.781172] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.781172] env[61824]: with excutils.save_and_reraise_exception(): [ 708.781172] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.781172] env[61824]: self.force_reraise() [ 708.781172] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.781172] env[61824]: raise self.value [ 708.781172] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.781172] env[61824]: updated_port = self._update_port( [ 708.781172] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.781172] env[61824]: _ensure_no_port_binding_failure(port) [ 708.781172] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.781172] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.782195] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. [ 708.782195] env[61824]: Removing descriptor: 17 [ 708.782195] env[61824]: ERROR nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Traceback (most recent call last): [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] yield resources [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self.driver.spawn(context, instance, image_meta, [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.782195] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] vm_ref = self.build_virtual_machine(instance, [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] for vif in network_info: [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return self._sync_wrapper(fn, *args, **kwargs) [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self.wait() [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self[:] = self._gt.wait() [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return self._exit_event.wait() [ 708.782585] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] result = hub.switch() [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return self.greenlet.switch() [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] result = function(*args, **kwargs) [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return func(*args, **kwargs) [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] raise e [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] nwinfo = self.network_api.allocate_for_instance( [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 708.782936] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] created_port_ids = self._update_ports_for_instance( [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] with excutils.save_and_reraise_exception(): [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self.force_reraise() [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] raise self.value [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] updated_port = self._update_port( [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] _ensure_no_port_binding_failure(port) [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.783401] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] raise exception.PortBindingFailed(port_id=port['id']) [ 708.783843] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] nova.exception.PortBindingFailed: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. [ 708.783843] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] [ 708.783843] env[61824]: INFO nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Terminating instance [ 708.783843] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.783843] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquired lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.783843] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.008547] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "0201e783-ed75-4cee-8df4-efac97b4ca19" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.008769] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.044445] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.044941] env[61824]: DEBUG nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.047429] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.308s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.126955] env[61824]: DEBUG oslo_concurrency.lockutils [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] Releasing lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.127230] env[61824]: DEBUG nova.compute.manager [req-39be1bef-ea15-4d28-9072-c2b3f413c109 req-769e31a9-06e4-4fb0-a071-a678a9506051 service nova] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Received event network-vif-deleted-5cb96a47-c415-423a-845e-fc3e1f820a35 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.303062] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.403752] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.554410] env[61824]: DEBUG nova.compute.utils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.560726] env[61824]: DEBUG nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Not allocating networking since 'none' was specified. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 709.896321] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336b6fca-e4da-4ad0-8b3f-e33635acbc04 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.904187] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea10a58-d646-4d30-83fa-d949229bc92d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.907560] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Releasing lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.907994] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.908207] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.908715] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-288e2c8e-259d-42c1-8b63-4010dfbd5e99 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.940636] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d324b2c4-3f02-4f09-bc10-e08fdf775aee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.946046] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1544554f-557b-422a-9e0a-ebbfc9181595 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.961563] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d323cee8-3baa-4fc4-a73f-2983fbc919c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.969082] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 662e815c-2548-4aed-a928-d98f9e1297e7 could not be found. [ 709.969082] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 709.969082] env[61824]: INFO nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Took 0.06 seconds to destroy the instance on the hypervisor. [ 709.969270] env[61824]: DEBUG oslo.service.loopingcall [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.969773] env[61824]: DEBUG nova.compute.manager [-] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.969869] env[61824]: DEBUG nova.network.neutron [-] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 709.979644] env[61824]: DEBUG nova.compute.provider_tree [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.994345] env[61824]: DEBUG nova.network.neutron [-] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.040566] env[61824]: DEBUG nova.compute.manager [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Received event network-changed-5c15fa5e-854c-4d05-922b-d032140c564c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.040766] env[61824]: DEBUG nova.compute.manager [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Refreshing instance network info cache due to event network-changed-5c15fa5e-854c-4d05-922b-d032140c564c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 710.040977] env[61824]: DEBUG oslo_concurrency.lockutils [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] Acquiring lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.041137] env[61824]: DEBUG oslo_concurrency.lockutils [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] Acquired lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.041294] env[61824]: DEBUG nova.network.neutron [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Refreshing network info cache for port 5c15fa5e-854c-4d05-922b-d032140c564c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.061236] env[61824]: DEBUG nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.482350] env[61824]: DEBUG nova.scheduler.client.report [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.497293] env[61824]: DEBUG nova.network.neutron [-] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.569246] env[61824]: DEBUG nova.network.neutron [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.673842] env[61824]: DEBUG nova.network.neutron [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.988026] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.940s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.988396] env[61824]: ERROR nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] Traceback (most recent call last): [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self.driver.spawn(context, instance, image_meta, [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] vm_ref = self.build_virtual_machine(instance, [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] vif_infos = vmwarevif.get_vif_info(self._session, [ 710.988396] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] for vif in network_info: [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return self._sync_wrapper(fn, *args, **kwargs) [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self.wait() [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self[:] = self._gt.wait() [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return self._exit_event.wait() [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] result = hub.switch() [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 710.988746] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return self.greenlet.switch() [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] result = function(*args, **kwargs) [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] return func(*args, **kwargs) [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] raise e [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] nwinfo = self.network_api.allocate_for_instance( [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] created_port_ids = self._update_ports_for_instance( [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] with excutils.save_and_reraise_exception(): [ 710.989182] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] self.force_reraise() [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] raise self.value [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] updated_port = self._update_port( [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] _ensure_no_port_binding_failure(port) [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] raise exception.PortBindingFailed(port_id=port['id']) [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] nova.exception.PortBindingFailed: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. [ 710.989582] env[61824]: ERROR nova.compute.manager [instance: 717eca7a-0287-4424-a302-5737f92724a6] [ 710.989939] env[61824]: DEBUG nova.compute.utils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 710.990345] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.189s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.993589] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Build of instance 717eca7a-0287-4424-a302-5737f92724a6 was re-scheduled: Binding failed for port f3e57e02-5736-4d6d-b52a-ba42f46d1689, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 710.993982] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 710.994215] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.994358] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquired lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.994533] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.000192] env[61824]: INFO nova.compute.manager [-] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Took 1.03 seconds to deallocate network for instance. [ 711.002264] env[61824]: DEBUG nova.compute.claims [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 711.002264] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.073228] env[61824]: DEBUG nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.097929] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.098189] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.098344] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.098520] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.098664] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.098809] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.099023] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.099184] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.099350] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.099508] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.099673] env[61824]: DEBUG nova.virt.hardware [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.100538] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db216f98-1410-4b09-922d-e271d7aa9af3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.109959] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1dfe3f-d755-4d89-8cac-7e3157d9de7e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.123094] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 711.128598] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Creating folder: Project (d1a8d9c3746748d2b0c65f5f60e7124d). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 711.128875] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35f9d455-0ddb-4641-960a-c5c1bb59e9b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.138870] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Created folder: Project (d1a8d9c3746748d2b0c65f5f60e7124d) in parent group-v274074. [ 711.139068] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Creating folder: Instances. Parent ref: group-v274094. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 711.139286] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-702b242d-e1cb-4582-afa1-0ba3a7c5e9fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.150049] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Created folder: Instances in parent group-v274094. [ 711.150274] env[61824]: DEBUG oslo.service.loopingcall [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.150457] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 711.150642] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0cb972c8-4ed6-4a59-ae3b-0f26d96506a4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.166144] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 711.166144] env[61824]: value = "task-1275368" [ 711.166144] env[61824]: _type = "Task" [ 711.166144] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.178305] env[61824]: DEBUG oslo_concurrency.lockutils [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] Releasing lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.178564] env[61824]: DEBUG nova.compute.manager [req-aeae97c4-af96-4ad7-b2ac-ea90bc002bf4 req-0e0fb760-0e17-475a-9d6a-2d67fb65c062 service nova] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Received event network-vif-deleted-5c15fa5e-854c-4d05-922b-d032140c564c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.178859] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275368, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.515042] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.589464] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.677602] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275368, 'name': CreateVM_Task, 'duration_secs': 0.239731} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.677786] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 711.678212] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.678368] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.678685] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 711.678934] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1878f1d-7c79-47c0-9092-096e648c2856 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.685106] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 711.685106] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52951d13-7cd3-f745-3ec1-7079911f41a6" [ 711.685106] env[61824]: _type = "Task" [ 711.685106] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.692620] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52951d13-7cd3-f745-3ec1-7079911f41a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.804413] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f256723c-db79-4612-90d2-cf6a65f3d1e3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.813315] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186aab36-ad8a-4b1d-8636-4173f142b893 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.843222] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80211681-2509-4ce8-bec2-de6177f83530 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.850308] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7ad043-3187-4f28-9288-cc636a699b7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.863112] env[61824]: DEBUG nova.compute.provider_tree [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.094428] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Releasing lock "refresh_cache-717eca7a-0287-4424-a302-5737f92724a6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.094768] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.094825] env[61824]: DEBUG nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.094987] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 712.114348] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.195881] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52951d13-7cd3-f745-3ec1-7079911f41a6, 'name': SearchDatastore_Task, 'duration_secs': 0.008428} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.196198] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.196421] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.196670] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.196823] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.196997] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.197255] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8da101d-2fab-40d5-9ebc-67d232275eab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.204540] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.204707] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 712.205388] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e1dba87-4c5b-4661-910c-660753dd155b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.209892] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 712.209892] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52076e01-b5a4-d567-a1a0-9c1558c6a1cc" [ 712.209892] env[61824]: _type = "Task" [ 712.209892] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.217044] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52076e01-b5a4-d567-a1a0-9c1558c6a1cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.366860] env[61824]: DEBUG nova.scheduler.client.report [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.618782] env[61824]: DEBUG nova.network.neutron [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.720108] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52076e01-b5a4-d567-a1a0-9c1558c6a1cc, 'name': SearchDatastore_Task, 'duration_secs': 0.007404} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.720834] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-014ab1c9-3d8f-4364-8073-25316df7ba00 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.726478] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 712.726478] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]524580db-6d64-1c34-ab1e-e0f1f911aeee" [ 712.726478] env[61824]: _type = "Task" [ 712.726478] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.733882] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524580db-6d64-1c34-ab1e-e0f1f911aeee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.872385] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.872385] env[61824]: ERROR nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. [ 712.872385] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Traceback (most recent call last): [ 712.872385] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.872385] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self.driver.spawn(context, instance, image_meta, [ 712.872385] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 712.872385] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.872385] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.872385] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] vm_ref = self.build_virtual_machine(instance, [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] for vif in network_info: [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return self._sync_wrapper(fn, *args, **kwargs) [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self.wait() [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self[:] = self._gt.wait() [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return self._exit_event.wait() [ 712.872850] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] result = hub.switch() [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return self.greenlet.switch() [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] result = function(*args, **kwargs) [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] return func(*args, **kwargs) [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] raise e [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] nwinfo = self.network_api.allocate_for_instance( [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 712.873337] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] created_port_ids = self._update_ports_for_instance( [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] with excutils.save_and_reraise_exception(): [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] self.force_reraise() [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] raise self.value [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] updated_port = self._update_port( [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] _ensure_no_port_binding_failure(port) [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.873741] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] raise exception.PortBindingFailed(port_id=port['id']) [ 712.874181] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] nova.exception.PortBindingFailed: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. [ 712.874181] env[61824]: ERROR nova.compute.manager [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] [ 712.874181] env[61824]: DEBUG nova.compute.utils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.874181] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.504s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.876951] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Build of instance 8aeeb8e9-e1ab-463b-bc16-ca68571882aa was re-scheduled: Binding failed for port 4ed391ee-ca92-4c4d-8cfc-04ed471a015a, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.877370] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.877583] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquiring lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.877750] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Acquired lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.877912] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 713.120530] env[61824]: INFO nova.compute.manager [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: 717eca7a-0287-4424-a302-5737f92724a6] Took 1.03 seconds to deallocate network for instance. [ 713.237025] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524580db-6d64-1c34-ab1e-e0f1f911aeee, 'name': SearchDatastore_Task, 'duration_secs': 0.008528} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.237245] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.237505] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 713.237798] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9744edc4-77fb-4980-ae2e-71ac6abd6c62 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.245544] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 713.245544] env[61824]: value = "task-1275369" [ 713.245544] env[61824]: _type = "Task" [ 713.245544] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.253999] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275369, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.405911] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.485431] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.744343] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7827e940-d134-4743-a8e4-a6ac0e6d48f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.757720] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e2e57f-3e8c-4bed-8258-34fcae7265c5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.760764] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275369, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466361} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.761016] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 713.761222] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 713.761740] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-968c3ee7-cb11-41e9-9820-ae5386216ddb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.789806] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6e9dc3-9ad1-46af-9645-b0dd54d12acb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.792347] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 713.792347] env[61824]: value = "task-1275370" [ 713.792347] env[61824]: _type = "Task" [ 713.792347] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.800871] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac9232e-493c-413a-abe4-079ee37d2b1e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.807969] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.818401] env[61824]: DEBUG nova.compute.provider_tree [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.988698] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Releasing lock "refresh_cache-8aeeb8e9-e1ab-463b-bc16-ca68571882aa" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.988854] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.989064] env[61824]: DEBUG nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.989250] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 714.004620] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.156974] env[61824]: INFO nova.scheduler.client.report [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Deleted allocations for instance 717eca7a-0287-4424-a302-5737f92724a6 [ 714.301418] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067457} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.301687] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.302420] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682f866d-1f15-4154-8c94-1b7eda1e3ee8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.321478] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.322274] env[61824]: DEBUG nova.scheduler.client.report [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.325131] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd853f9e-6ac5-4caa-b6ad-4189ee4628ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.339730] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.466s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.340300] env[61824]: ERROR nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Traceback (most recent call last): [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self.driver.spawn(context, instance, image_meta, [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] vm_ref = self.build_virtual_machine(instance, [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.340300] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] for vif in network_info: [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return self._sync_wrapper(fn, *args, **kwargs) [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self.wait() [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self[:] = self._gt.wait() [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return self._exit_event.wait() [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] result = hub.switch() [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 714.340705] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return self.greenlet.switch() [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] result = function(*args, **kwargs) [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] return func(*args, **kwargs) [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] raise e [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] nwinfo = self.network_api.allocate_for_instance( [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] created_port_ids = self._update_ports_for_instance( [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] with excutils.save_and_reraise_exception(): [ 714.341130] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] self.force_reraise() [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] raise self.value [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] updated_port = self._update_port( [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] _ensure_no_port_binding_failure(port) [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] raise exception.PortBindingFailed(port_id=port['id']) [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] nova.exception.PortBindingFailed: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. [ 714.341538] env[61824]: ERROR nova.compute.manager [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] [ 714.341925] env[61824]: DEBUG nova.compute.utils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 714.341925] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.877s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.343533] env[61824]: INFO nova.compute.claims [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.346139] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Build of instance fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e was re-scheduled: Binding failed for port fece6a3b-5cd0-4307-aa22-a2b0e30ccc6f, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 714.346515] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 714.346756] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquiring lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.346907] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Acquired lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.347077] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.354342] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 714.354342] env[61824]: value = "task-1275371" [ 714.354342] env[61824]: _type = "Task" [ 714.354342] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.361906] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275371, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.507384] env[61824]: DEBUG nova.network.neutron [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.665362] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ea066262-0892-496e-a8dd-afcc63aca5dd tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "717eca7a-0287-4424-a302-5737f92724a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.382s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.864100] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275371, 'name': ReconfigVM_Task, 'duration_secs': 0.30526} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.864384] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Reconfigured VM instance instance-0000002a to attach disk [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 714.864995] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4d002b0b-31d4-474f-9be1-e64730c6f50e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.871350] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 714.871350] env[61824]: value = "task-1275372" [ 714.871350] env[61824]: _type = "Task" [ 714.871350] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.874667] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.881721] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275372, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.953922] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.010496] env[61824]: INFO nova.compute.manager [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] [instance: 8aeeb8e9-e1ab-463b-bc16-ca68571882aa] Took 1.02 seconds to deallocate network for instance. [ 715.168477] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.382857] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275372, 'name': Rename_Task, 'duration_secs': 0.155082} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.383251] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 715.385373] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f04d05d-5a24-4914-8268-d4c66bbc6d26 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.391063] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 715.391063] env[61824]: value = "task-1275373" [ 715.391063] env[61824]: _type = "Task" [ 715.391063] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.398856] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.457118] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Releasing lock "refresh_cache-fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.457424] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 715.457605] env[61824]: DEBUG nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 715.457707] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 715.474399] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.676462] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf47595-e0a8-487d-99dd-60e64ea51d09 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.687638] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddb7246-4f49-4134-8793-6b4dc77ed65b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.692027] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.718495] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78f6c2f-4eb0-42be-aa25-3c4689a8920b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.727311] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330dac2f-45b6-4979-97cd-ea3543ef2765 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.742858] env[61824]: DEBUG nova.compute.provider_tree [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.902248] env[61824]: DEBUG oslo_vmware.api [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275373, 'name': PowerOnVM_Task, 'duration_secs': 0.415208} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.902698] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 715.903034] env[61824]: INFO nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Took 4.83 seconds to spawn the instance on the hypervisor. [ 715.903347] env[61824]: DEBUG nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 715.904244] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f49851c-d27d-4326-a9d4-19597e0d2a06 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.976711] env[61824]: DEBUG nova.network.neutron [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.039968] env[61824]: INFO nova.scheduler.client.report [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Deleted allocations for instance 8aeeb8e9-e1ab-463b-bc16-ca68571882aa [ 716.246059] env[61824]: DEBUG nova.scheduler.client.report [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.420551] env[61824]: INFO nova.compute.manager [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Took 26.50 seconds to build instance. [ 716.479615] env[61824]: INFO nova.compute.manager [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] [instance: fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e] Took 1.02 seconds to deallocate network for instance. [ 716.548660] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a8ee6008-d00f-435f-b9b4-a8c769e0d0b9 tempest-MigrationsAdminTest-1277074355 tempest-MigrationsAdminTest-1277074355-project-member] Lock "8aeeb8e9-e1ab-463b-bc16-ca68571882aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.912s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.751870] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.752423] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.755773] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.323s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.757551] env[61824]: INFO nova.compute.claims [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.922357] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5ce9a501-7e83-433f-a757-bac7fac155cb tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "e758c89f-e5e2-42e0-926e-c629cb496a89" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.089s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.051093] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.265500] env[61824]: DEBUG nova.compute.utils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.266726] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.266894] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 717.322441] env[61824]: DEBUG nova.policy [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1b131826ac644a639a28c4db774bee92', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4ad43bb71692464ea67ffe9519858f51', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.388199] env[61824]: INFO nova.compute.manager [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Rebuilding instance [ 717.425359] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.432838] env[61824]: DEBUG nova.compute.manager [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 717.433706] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2fdfd4-b768-4e32-ade2-7e49377a62ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.510296] env[61824]: INFO nova.scheduler.client.report [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Deleted allocations for instance fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e [ 717.571872] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.636029] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Successfully created port: 14a7d012-82d8-4fed-a618-cd3336d5b57e {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.774125] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.944956] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 717.947563] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.947864] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bd4b81a-b889-411a-b7dd-6c73823aec48 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.960018] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 717.960018] env[61824]: value = "task-1275374" [ 717.960018] env[61824]: _type = "Task" [ 717.960018] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.970808] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275374, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.015786] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5527afcb-2c0b-43ba-a6c7-3d539157eac5 tempest-ListImageFiltersTestJSON-784476940 tempest-ListImageFiltersTestJSON-784476940-project-member] Lock "fb4c4bbf-ca9c-4584-bd1d-83e5e81e6c7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.018s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.170940] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f648d57a-702e-4c76-904e-3f372bec1208 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.179108] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11fd066-3b36-4e79-b3d7-3400560f0d5d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.208431] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0298c7-eb33-4f75-ba58-61561affa0c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.220376] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e48140-bebf-47b5-b2da-ae2a63c31356 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.235209] env[61824]: DEBUG nova.compute.provider_tree [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.468996] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275374, 'name': PowerOffVM_Task, 'duration_secs': 0.189517} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.468996] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 718.468996] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.469649] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c85953-c0f3-4605-a4e9-592fafde4d45 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.476056] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 718.476295] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b86255c-cbfc-44d3-ab66-606c360e07d2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.500705] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 718.500932] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 718.501122] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Deleting the datastore file [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 718.501397] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1787529-8178-4003-bc6c-c85f9f9bc577 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.507519] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 718.507519] env[61824]: value = "task-1275376" [ 718.507519] env[61824]: _type = "Task" [ 718.507519] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.515235] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275376, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.521724] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 718.615431] env[61824]: DEBUG nova.compute.manager [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Received event network-changed-14a7d012-82d8-4fed-a618-cd3336d5b57e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.615897] env[61824]: DEBUG nova.compute.manager [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Refreshing instance network info cache due to event network-changed-14a7d012-82d8-4fed-a618-cd3336d5b57e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.616283] env[61824]: DEBUG oslo_concurrency.lockutils [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] Acquiring lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.616574] env[61824]: DEBUG oslo_concurrency.lockutils [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] Acquired lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.618079] env[61824]: DEBUG nova.network.neutron [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Refreshing network info cache for port 14a7d012-82d8-4fed-a618-cd3336d5b57e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 718.738300] env[61824]: DEBUG nova.scheduler.client.report [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.784816] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.818152] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.818395] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.818551] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.818725] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.818863] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.819021] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.819795] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.819970] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.820156] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.820322] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.821680] env[61824]: DEBUG nova.virt.hardware [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.822706] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603dd29f-924a-4575-91be-29527a1dbf11 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.827478] env[61824]: ERROR nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. [ 718.827478] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.827478] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.827478] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.827478] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 718.827478] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.827478] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 718.827478] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.827478] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.827478] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 718.827478] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.827478] env[61824]: ERROR nova.compute.manager raise self.value [ 718.827478] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 718.827478] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.827478] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.827478] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.827957] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.827957] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.827957] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. [ 718.827957] env[61824]: ERROR nova.compute.manager [ 718.827957] env[61824]: Traceback (most recent call last): [ 718.827957] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.827957] env[61824]: listener.cb(fileno) [ 718.827957] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.827957] env[61824]: result = function(*args, **kwargs) [ 718.827957] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.827957] env[61824]: return func(*args, **kwargs) [ 718.827957] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.827957] env[61824]: raise e [ 718.827957] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.827957] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 718.827957] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 718.827957] env[61824]: created_port_ids = self._update_ports_for_instance( [ 718.827957] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 718.827957] env[61824]: with excutils.save_and_reraise_exception(): [ 718.827957] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.827957] env[61824]: self.force_reraise() [ 718.827957] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.827957] env[61824]: raise self.value [ 718.827957] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 718.827957] env[61824]: updated_port = self._update_port( [ 718.827957] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.827957] env[61824]: _ensure_no_port_binding_failure(port) [ 718.827957] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.827957] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.828816] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. [ 718.828816] env[61824]: Removing descriptor: 17 [ 718.835037] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1a2e94-efa0-4534-835f-a45687abfa14 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.849113] env[61824]: ERROR nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Traceback (most recent call last): [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] yield resources [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self.driver.spawn(context, instance, image_meta, [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] vm_ref = self.build_virtual_machine(instance, [ 718.849113] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] for vif in network_info: [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] return self._sync_wrapper(fn, *args, **kwargs) [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self.wait() [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self[:] = self._gt.wait() [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] return self._exit_event.wait() [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 718.849499] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] current.throw(*self._exc) [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] result = function(*args, **kwargs) [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] return func(*args, **kwargs) [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] raise e [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] nwinfo = self.network_api.allocate_for_instance( [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] created_port_ids = self._update_ports_for_instance( [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] with excutils.save_and_reraise_exception(): [ 718.849851] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self.force_reraise() [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] raise self.value [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] updated_port = self._update_port( [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] _ensure_no_port_binding_failure(port) [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] raise exception.PortBindingFailed(port_id=port['id']) [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] nova.exception.PortBindingFailed: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. [ 718.850243] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] [ 718.850243] env[61824]: INFO nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Terminating instance [ 718.851795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.017241] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275376, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089146} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.017539] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 719.017670] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 719.017938] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.042420] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.142776] env[61824]: DEBUG nova.network.neutron [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.244368] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.246190] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.248128] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.203s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.251872] env[61824]: DEBUG nova.network.neutron [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.752295] env[61824]: DEBUG nova.compute.utils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.754575] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.754856] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.759685] env[61824]: DEBUG oslo_concurrency.lockutils [req-2ae84a91-310f-4381-b56c-877f3f1d1f98 req-5c99f641-f4fc-4865-ac23-1e5f57af4bd7 service nova] Releasing lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.760240] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquired lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.760506] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 719.821447] env[61824]: DEBUG nova.policy [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e7fc746334a4b7a9153a721b6eafcfb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c09c426f3fd349c190e1058f8ec87634', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.059770] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.060070] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.060266] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.060484] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.060656] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.060835] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.061097] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.061296] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.061497] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.061693] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.061897] env[61824]: DEBUG nova.virt.hardware [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.063182] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322cb553-f007-4354-ad43-c79a22bef9ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.071463] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dab58a2-b99a-4893-ad56-2641dcf72e6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.089023] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 720.095377] env[61824]: DEBUG oslo.service.loopingcall [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.098110] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 720.098513] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3e62168-b8fe-4c27-8f0d-c5483668779f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.124031] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 720.124031] env[61824]: value = "task-1275377" [ 720.124031] env[61824]: _type = "Task" [ 720.124031] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.133369] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275377, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.170945] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bf6b17-529b-4410-bb4e-61c91e2c2c80 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.179565] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30bdd034-7ad5-461c-be80-9691f8f0258d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.214276] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28d16ae-9c76-4507-8010-0d6374059557 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.224013] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae418f3-d68a-43bf-9438-ef3e3962f90a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.241218] env[61824]: DEBUG nova.compute.provider_tree [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.242682] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Successfully created port: 256f6905-aca8-42b9-8586-32db2f6d3f5b {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.258110] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.281645] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 720.373664] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.634404] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275377, 'name': CreateVM_Task, 'duration_secs': 0.246637} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.634404] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 720.634404] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.634404] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.634913] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 720.634913] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06828dc8-f720-4124-82bc-e4a38f5fc233 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.639248] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 720.639248] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52213274-8694-2e04-970f-e1df77933f23" [ 720.639248] env[61824]: _type = "Task" [ 720.639248] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.646915] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52213274-8694-2e04-970f-e1df77933f23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.708181] env[61824]: DEBUG nova.compute.manager [req-3f27bd4b-554a-4148-9533-049b04c7ab28 req-ef099cdc-a5bc-48e2-8a07-5c7a199b5d73 service nova] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Received event network-vif-deleted-14a7d012-82d8-4fed-a618-cd3336d5b57e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.750208] env[61824]: DEBUG nova.scheduler.client.report [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.879818] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Releasing lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.880265] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.883114] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.883114] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8f52833-8a12-45af-9a7d-b89d91b256e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.891654] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77d6d1d-9014-48f6-a43b-b464e59af870 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.921411] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 204fd4be-4a16-4838-a30d-c78339b3834d could not be found. [ 720.921411] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.921411] env[61824]: INFO nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 720.921411] env[61824]: DEBUG oslo.service.loopingcall [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.921411] env[61824]: DEBUG nova.compute.manager [-] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.921411] env[61824]: DEBUG nova.network.neutron [-] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 720.936948] env[61824]: DEBUG nova.network.neutron [-] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.155340] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52213274-8694-2e04-970f-e1df77933f23, 'name': SearchDatastore_Task, 'duration_secs': 0.009738} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.155743] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.156113] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 721.156396] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.156582] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.156842] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 721.157185] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc37dd5c-4971-458c-8be7-669d589a7254 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.165300] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 721.165518] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 721.167944] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f1566db-d2e2-427d-81b3-0613ff4a5045 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.173643] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 721.173643] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52da436f-32fb-70eb-28cc-e1f1379a12d2" [ 721.173643] env[61824]: _type = "Task" [ 721.173643] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.184344] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52da436f-32fb-70eb-28cc-e1f1379a12d2, 'name': SearchDatastore_Task, 'duration_secs': 0.007827} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.185136] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ee05a35-1ba8-44cc-bf86-7d9371b977e4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.189765] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 721.189765] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]523fa544-01d8-995e-8b8d-d428f58dc862" [ 721.189765] env[61824]: _type = "Task" [ 721.189765] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.204840] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523fa544-01d8-995e-8b8d-d428f58dc862, 'name': SearchDatastore_Task, 'duration_secs': 0.007665} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.205428] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.206669] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 721.206983] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-852d0719-10a1-4e19-b08c-b4fce5381d64 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.213682] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 721.213682] env[61824]: value = "task-1275378" [ 721.213682] env[61824]: _type = "Task" [ 721.213682] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.223166] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275378, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.257442] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.009s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.258135] env[61824]: ERROR nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Traceback (most recent call last): [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self.driver.spawn(context, instance, image_meta, [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] vm_ref = self.build_virtual_machine(instance, [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.258135] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] for vif in network_info: [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return self._sync_wrapper(fn, *args, **kwargs) [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self.wait() [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self[:] = self._gt.wait() [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return self._exit_event.wait() [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] result = hub.switch() [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.258616] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return self.greenlet.switch() [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] result = function(*args, **kwargs) [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] return func(*args, **kwargs) [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] raise e [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] nwinfo = self.network_api.allocate_for_instance( [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] created_port_ids = self._update_ports_for_instance( [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] with excutils.save_and_reraise_exception(): [ 721.259123] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] self.force_reraise() [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] raise self.value [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] updated_port = self._update_port( [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] _ensure_no_port_binding_failure(port) [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] raise exception.PortBindingFailed(port_id=port['id']) [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] nova.exception.PortBindingFailed: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. [ 721.259553] env[61824]: ERROR nova.compute.manager [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] [ 721.260026] env[61824]: DEBUG nova.compute.utils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.260665] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.010s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.270252] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Build of instance 037dd0ba-0d50-4c46-8947-7df3ca19e1c1 was re-scheduled: Binding failed for port 1c4e8983-546f-4c22-98cd-5f12b4c8322d, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.270252] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.270252] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.270252] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.270469] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.270469] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 721.305616] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 721.305853] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 721.306021] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 721.306203] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 721.306347] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 721.306490] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 721.306707] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 721.306887] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 721.307065] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 721.307228] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 721.307395] env[61824]: DEBUG nova.virt.hardware [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 721.308329] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0617f7de-0240-4e7c-9f26-b64cf5843ec5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.318212] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a341b1f7-ed17-4186-9f16-4f7eeb0e8dba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.439765] env[61824]: DEBUG nova.network.neutron [-] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.692118] env[61824]: ERROR nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. [ 721.692118] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 721.692118] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.692118] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 721.692118] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.692118] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 721.692118] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.692118] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 721.692118] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.692118] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 721.692118] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.692118] env[61824]: ERROR nova.compute.manager raise self.value [ 721.692118] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.692118] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 721.692118] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.692118] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 721.692728] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.692728] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 721.692728] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. [ 721.692728] env[61824]: ERROR nova.compute.manager [ 721.692728] env[61824]: Traceback (most recent call last): [ 721.692728] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 721.692728] env[61824]: listener.cb(fileno) [ 721.692728] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.692728] env[61824]: result = function(*args, **kwargs) [ 721.692728] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.692728] env[61824]: return func(*args, **kwargs) [ 721.692728] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.692728] env[61824]: raise e [ 721.692728] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.692728] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 721.692728] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.692728] env[61824]: created_port_ids = self._update_ports_for_instance( [ 721.692728] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.692728] env[61824]: with excutils.save_and_reraise_exception(): [ 721.692728] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.692728] env[61824]: self.force_reraise() [ 721.692728] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.692728] env[61824]: raise self.value [ 721.692728] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.692728] env[61824]: updated_port = self._update_port( [ 721.692728] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.692728] env[61824]: _ensure_no_port_binding_failure(port) [ 721.692728] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.692728] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 721.693644] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. [ 721.693644] env[61824]: Removing descriptor: 17 [ 721.693644] env[61824]: ERROR nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] Traceback (most recent call last): [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] yield resources [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self.driver.spawn(context, instance, image_meta, [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.693644] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] vm_ref = self.build_virtual_machine(instance, [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] for vif in network_info: [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return self._sync_wrapper(fn, *args, **kwargs) [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self.wait() [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self[:] = self._gt.wait() [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return self._exit_event.wait() [ 721.694106] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] result = hub.switch() [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return self.greenlet.switch() [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] result = function(*args, **kwargs) [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return func(*args, **kwargs) [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] raise e [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] nwinfo = self.network_api.allocate_for_instance( [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 721.694580] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] created_port_ids = self._update_ports_for_instance( [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] with excutils.save_and_reraise_exception(): [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self.force_reraise() [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] raise self.value [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] updated_port = self._update_port( [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] _ensure_no_port_binding_failure(port) [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.695066] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] raise exception.PortBindingFailed(port_id=port['id']) [ 721.695450] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] nova.exception.PortBindingFailed: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. [ 721.695450] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] [ 721.695450] env[61824]: INFO nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Terminating instance [ 721.695450] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.695591] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquired lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.695628] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.724476] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275378, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458305} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.724744] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 721.724961] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.725263] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b23e305d-c2b1-442d-bf3e-b42bc2b052e5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.731253] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 721.731253] env[61824]: value = "task-1275379" [ 721.731253] env[61824]: _type = "Task" [ 721.731253] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.739582] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275379, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.795640] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.873221] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.943983] env[61824]: INFO nova.compute.manager [-] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Took 1.02 seconds to deallocate network for instance. [ 721.946310] env[61824]: DEBUG nova.compute.claims [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.946479] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.064425] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d329079-86d5-4b45-a369-dc1987055209 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.071874] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5b23e0-0bb5-4fac-9c76-30b381da7591 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.104333] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6637c85-de28-450e-94b7-bbbd98edd2c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.111538] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f06961bc-1ceb-49d6-ad1e-267e002a2c05 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.124905] env[61824]: DEBUG nova.compute.provider_tree [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.219423] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.240868] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275379, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064661} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.241142] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.241890] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35651209-b02b-4cfe-b33a-3a35f6ced6b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.261014] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.263247] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1abc7a4b-0e42-4e55-aa8a-8d762aad665f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.283783] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 722.283783] env[61824]: value = "task-1275380" [ 722.283783] env[61824]: _type = "Task" [ 722.283783] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.293172] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275380, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.320827] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.376861] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "refresh_cache-037dd0ba-0d50-4c46-8947-7df3ca19e1c1" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.377200] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.377448] env[61824]: DEBUG nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.377671] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 722.394629] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.628687] env[61824]: DEBUG nova.scheduler.client.report [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.731428] env[61824]: DEBUG nova.compute.manager [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Received event network-changed-256f6905-aca8-42b9-8586-32db2f6d3f5b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.731650] env[61824]: DEBUG nova.compute.manager [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Refreshing instance network info cache due to event network-changed-256f6905-aca8-42b9-8586-32db2f6d3f5b. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.731806] env[61824]: DEBUG oslo_concurrency.lockutils [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] Acquiring lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.795306] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275380, 'name': ReconfigVM_Task, 'duration_secs': 0.251055} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.795587] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Reconfigured VM instance instance-0000002a to attach disk [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89/e758c89f-e5e2-42e0-926e-c629cb496a89.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 722.796448] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f34a91e-314e-4a23-9583-f040519a0fb7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.802461] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 722.802461] env[61824]: value = "task-1275381" [ 722.802461] env[61824]: _type = "Task" [ 722.802461] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.809499] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275381, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.823025] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Releasing lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.823402] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.823582] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.823839] env[61824]: DEBUG oslo_concurrency.lockutils [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] Acquired lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.824013] env[61824]: DEBUG nova.network.neutron [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Refreshing network info cache for port 256f6905-aca8-42b9-8586-32db2f6d3f5b {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 722.824971] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7618ccf3-b3c6-4d0b-b113-e546e49c868f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.833517] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3ba86a-6e0a-4c82-aaca-ec195188dc5e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.855145] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f525feb8-d24a-4f89-b640-471163d181ef could not be found. [ 722.855341] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.855510] env[61824]: INFO nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Took 0.03 seconds to destroy the instance on the hypervisor. [ 722.855755] env[61824]: DEBUG oslo.service.loopingcall [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.855983] env[61824]: DEBUG nova.compute.manager [-] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.856090] env[61824]: DEBUG nova.network.neutron [-] [instance: f525feb8-d24a-4f89-b640-471163d181ef] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 722.871185] env[61824]: DEBUG nova.network.neutron [-] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.896964] env[61824]: DEBUG nova.network.neutron [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.132947] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.873s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.133608] env[61824]: ERROR nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Traceback (most recent call last): [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self.driver.spawn(context, instance, image_meta, [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] vm_ref = self.build_virtual_machine(instance, [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.133608] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] for vif in network_info: [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return self._sync_wrapper(fn, *args, **kwargs) [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self.wait() [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self[:] = self._gt.wait() [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return self._exit_event.wait() [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] result = hub.switch() [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.133982] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return self.greenlet.switch() [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] result = function(*args, **kwargs) [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] return func(*args, **kwargs) [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] raise e [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] nwinfo = self.network_api.allocate_for_instance( [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] created_port_ids = self._update_ports_for_instance( [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] with excutils.save_and_reraise_exception(): [ 723.134406] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] self.force_reraise() [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] raise self.value [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] updated_port = self._update_port( [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] _ensure_no_port_binding_failure(port) [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] raise exception.PortBindingFailed(port_id=port['id']) [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] nova.exception.PortBindingFailed: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. [ 723.134964] env[61824]: ERROR nova.compute.manager [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] [ 723.135311] env[61824]: DEBUG nova.compute.utils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.136040] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.452s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.138846] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Build of instance 6e2190ac-d287-4810-9d43-740e62ae7b56 was re-scheduled: Binding failed for port caf0c180-8fb5-411d-b711-0075a7403c83, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 723.139634] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 723.139860] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Acquiring lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.140134] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Acquired lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.140310] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.313192] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275381, 'name': Rename_Task, 'duration_secs': 0.132081} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.313192] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 723.313412] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92a0c402-b14b-4d7b-b48e-0fdcea75137d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.320084] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 723.320084] env[61824]: value = "task-1275382" [ 723.320084] env[61824]: _type = "Task" [ 723.320084] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.327152] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275382, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.346668] env[61824]: DEBUG nova.network.neutron [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.373172] env[61824]: DEBUG nova.network.neutron [-] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.399968] env[61824]: INFO nova.compute.manager [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 037dd0ba-0d50-4c46-8947-7df3ca19e1c1] Took 1.02 seconds to deallocate network for instance. [ 723.453807] env[61824]: DEBUG nova.network.neutron [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.674354] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.775047] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.831328] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275382, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.877028] env[61824]: INFO nova.compute.manager [-] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Took 1.02 seconds to deallocate network for instance. [ 723.878645] env[61824]: DEBUG nova.compute.claims [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.878819] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.945980] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06184809-9e30-488a-a60d-25adc54b6ca4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.953950] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb7dcbf-6666-4eaf-a9ef-563633bc7bed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.957133] env[61824]: DEBUG oslo_concurrency.lockutils [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] Releasing lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.958065] env[61824]: DEBUG nova.compute.manager [req-7ed52cd6-a53c-4d79-8e5d-9afe462f6734 req-1f7cf9e4-e264-42a1-a3a1-184a0dc381b6 service nova] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Received event network-vif-deleted-256f6905-aca8-42b9-8586-32db2f6d3f5b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.984715] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd04bd1-c599-46c6-8da0-a4192b84fdf3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.992169] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aff0100-c0f2-4ed7-9bc3-83c00a0b3da2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.006047] env[61824]: DEBUG nova.compute.provider_tree [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.278604] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Releasing lock "refresh_cache-6e2190ac-d287-4810-9d43-740e62ae7b56" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.278976] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 724.278976] env[61824]: DEBUG nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.279160] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.294075] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.331040] env[61824]: DEBUG oslo_vmware.api [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275382, 'name': PowerOnVM_Task, 'duration_secs': 0.955168} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.331040] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 724.331040] env[61824]: DEBUG nova.compute.manager [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 724.331275] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5106001d-fe78-43c0-acc1-5388a34b28a3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.385549] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.386188] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.431263] env[61824]: INFO nova.scheduler.client.report [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Deleted allocations for instance 037dd0ba-0d50-4c46-8947-7df3ca19e1c1 [ 724.508619] env[61824]: DEBUG nova.scheduler.client.report [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.797093] env[61824]: DEBUG nova.network.neutron [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.845589] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.891778] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 724.891936] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 724.892068] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Rebuilding the list of instances to heal {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 724.938733] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fcd4b7f3-40ad-45c8-98da-b8104c235919 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "037dd0ba-0d50-4c46-8947-7df3ca19e1c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.915s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.013727] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.878s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.014388] env[61824]: ERROR nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Traceback (most recent call last): [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self.driver.spawn(context, instance, image_meta, [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] vm_ref = self.build_virtual_machine(instance, [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.014388] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] for vif in network_info: [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return self._sync_wrapper(fn, *args, **kwargs) [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self.wait() [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self[:] = self._gt.wait() [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return self._exit_event.wait() [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] result = hub.switch() [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 725.014836] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return self.greenlet.switch() [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] result = function(*args, **kwargs) [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] return func(*args, **kwargs) [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] raise e [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] nwinfo = self.network_api.allocate_for_instance( [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] created_port_ids = self._update_ports_for_instance( [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] with excutils.save_and_reraise_exception(): [ 725.015311] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] self.force_reraise() [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] raise self.value [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] updated_port = self._update_port( [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] _ensure_no_port_binding_failure(port) [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] raise exception.PortBindingFailed(port_id=port['id']) [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] nova.exception.PortBindingFailed: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. [ 725.015725] env[61824]: ERROR nova.compute.manager [instance: dc662f50-f111-4d26-b017-3ab719004ac2] [ 725.016128] env[61824]: DEBUG nova.compute.utils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.016503] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.014s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.019574] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Build of instance dc662f50-f111-4d26-b017-3ab719004ac2 was re-scheduled: Binding failed for port 5cb96a47-c415-423a-845e-fc3e1f820a35, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.019989] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.020224] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.020368] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquired lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.020523] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.299640] env[61824]: INFO nova.compute.manager [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] [instance: 6e2190ac-d287-4810-9d43-740e62ae7b56] Took 1.02 seconds to deallocate network for instance. [ 725.398108] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 725.398307] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 725.398436] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 725.416712] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "refresh_cache-e758c89f-e5e2-42e0-926e-c629cb496a89" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.416891] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquired lock "refresh_cache-e758c89f-e5e2-42e0-926e-c629cb496a89" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.417057] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Forcefully refreshing network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 725.417214] env[61824]: DEBUG nova.objects.instance [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lazy-loading 'info_cache' on Instance uuid e758c89f-e5e2-42e0-926e-c629cb496a89 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 725.433909] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "e758c89f-e5e2-42e0-926e-c629cb496a89" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.434150] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "e758c89f-e5e2-42e0-926e-c629cb496a89" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.435693] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "e758c89f-e5e2-42e0-926e-c629cb496a89-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.435693] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "e758c89f-e5e2-42e0-926e-c629cb496a89-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.435693] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "e758c89f-e5e2-42e0-926e-c629cb496a89-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.436941] env[61824]: INFO nova.compute.manager [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Terminating instance [ 725.438258] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "refresh_cache-e758c89f-e5e2-42e0-926e-c629cb496a89" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.440876] env[61824]: DEBUG nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.539558] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.629129] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.790983] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8085feba-84ea-46c9-b35e-31d652a3dee8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.799299] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d3706e-e04b-4116-b35a-1a16545bb864 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.834915] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e50b74b-7a82-46ac-a02e-62c63f38dce1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.842773] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c94a2f-e4c4-4030-9963-35ef843a9067 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.856766] env[61824]: DEBUG nova.compute.provider_tree [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.960272] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.133659] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Releasing lock "refresh_cache-dc662f50-f111-4d26-b017-3ab719004ac2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.133659] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 726.133659] env[61824]: DEBUG nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.133659] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.162028] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.335400] env[61824]: INFO nova.scheduler.client.report [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Deleted allocations for instance 6e2190ac-d287-4810-9d43-740e62ae7b56 [ 726.359523] env[61824]: DEBUG nova.scheduler.client.report [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.438535] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.667359] env[61824]: DEBUG nova.network.neutron [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.845081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1cdc815c-5f7b-453a-bf67-b45048c2e462 tempest-ServersTestFqdnHostnames-922121389 tempest-ServersTestFqdnHostnames-922121389-project-member] Lock "6e2190ac-d287-4810-9d43-740e62ae7b56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.198s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.865872] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.849s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.866263] env[61824]: ERROR nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Traceback (most recent call last): [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self.driver.spawn(context, instance, image_meta, [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] vm_ref = self.build_virtual_machine(instance, [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.866263] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] for vif in network_info: [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return self._sync_wrapper(fn, *args, **kwargs) [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self.wait() [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self[:] = self._gt.wait() [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return self._exit_event.wait() [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] result = hub.switch() [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.866597] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return self.greenlet.switch() [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] result = function(*args, **kwargs) [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] return func(*args, **kwargs) [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] raise e [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] nwinfo = self.network_api.allocate_for_instance( [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] created_port_ids = self._update_ports_for_instance( [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] with excutils.save_and_reraise_exception(): [ 726.867068] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] self.force_reraise() [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] raise self.value [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] updated_port = self._update_port( [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] _ensure_no_port_binding_failure(port) [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] raise exception.PortBindingFailed(port_id=port['id']) [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] nova.exception.PortBindingFailed: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. [ 726.867464] env[61824]: ERROR nova.compute.manager [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] [ 726.867914] env[61824]: DEBUG nova.compute.utils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.868465] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.177s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.870470] env[61824]: INFO nova.compute.claims [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.874284] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Build of instance 662e815c-2548-4aed-a928-d98f9e1297e7 was re-scheduled: Binding failed for port 5c15fa5e-854c-4d05-922b-d032140c564c, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 726.874284] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 726.874284] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.874284] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquired lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.874525] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.021180] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.170767] env[61824]: INFO nova.compute.manager [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: dc662f50-f111-4d26-b017-3ab719004ac2] Took 1.04 seconds to deallocate network for instance. [ 727.348146] env[61824]: DEBUG nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.405159] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.446658] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "d9bb3172-c058-4e59-a719-d8aa340fba71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.447071] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.501029] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.526179] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Releasing lock "refresh_cache-e758c89f-e5e2-42e0-926e-c629cb496a89" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.526179] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Updated the network info_cache for instance {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 727.526179] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquired lock "refresh_cache-e758c89f-e5e2-42e0-926e-c629cb496a89" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.526179] env[61824]: DEBUG nova.network.neutron [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.526390] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 727.526573] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 727.526955] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 727.527130] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 727.527512] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 727.527674] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 727.527827] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 727.527983] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 727.874416] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.004497] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Releasing lock "refresh_cache-662e815c-2548-4aed-a928-d98f9e1297e7" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.004723] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 728.004907] env[61824]: DEBUG nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.005742] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 728.026427] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.032105] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.045186] env[61824]: DEBUG nova.network.neutron [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.160449] env[61824]: DEBUG nova.network.neutron [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.198836] env[61824]: INFO nova.scheduler.client.report [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Deleted allocations for instance dc662f50-f111-4d26-b017-3ab719004ac2 [ 728.208017] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3572d1bb-314d-45f1-94f6-800cc9979cb2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.215182] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71c5d79-d82c-409e-a78b-80ce144534f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.249943] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc17fd6-3a38-4de0-9bc1-1f9172679b71 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.258308] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fc1fe1-0b26-4181-8a70-9fcd13abcf6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.275066] env[61824]: DEBUG nova.compute.provider_tree [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.536016] env[61824]: DEBUG nova.network.neutron [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.664786] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Releasing lock "refresh_cache-e758c89f-e5e2-42e0-926e-c629cb496a89" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.665286] env[61824]: DEBUG nova.compute.manager [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.665487] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.666386] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb80083-98be-4db7-8a25-f2333df48d28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.683331] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 728.683589] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d29d606-3928-4228-b773-1c51c3067833 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.690787] env[61824]: DEBUG oslo_vmware.api [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 728.690787] env[61824]: value = "task-1275383" [ 728.690787] env[61824]: _type = "Task" [ 728.690787] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.702462] env[61824]: DEBUG oslo_vmware.api [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.707013] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f597f585-f23a-4df8-9ed2-5589523e011b tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "dc662f50-f111-4d26-b017-3ab719004ac2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.913s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.778708] env[61824]: DEBUG nova.scheduler.client.report [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.036956] env[61824]: INFO nova.compute.manager [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 662e815c-2548-4aed-a928-d98f9e1297e7] Took 1.03 seconds to deallocate network for instance. [ 729.206767] env[61824]: DEBUG oslo_vmware.api [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275383, 'name': PowerOffVM_Task, 'duration_secs': 0.193292} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.206767] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 729.206767] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 729.206767] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-779acba6-92ce-48f9-b185-66144b0487f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.210241] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.230056] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 729.230274] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 729.230476] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Deleting the datastore file [datastore2] e758c89f-e5e2-42e0-926e-c629cb496a89 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 729.231395] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5eb51723-f1fc-48f6-8908-c257620e7185 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.237068] env[61824]: DEBUG oslo_vmware.api [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for the task: (returnval){ [ 729.237068] env[61824]: value = "task-1275385" [ 729.237068] env[61824]: _type = "Task" [ 729.237068] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.245624] env[61824]: DEBUG oslo_vmware.api [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275385, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.286174] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.286174] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 729.287896] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.716s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.290418] env[61824]: INFO nova.compute.claims [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.733553] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.746704] env[61824]: DEBUG oslo_vmware.api [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Task: {'id': task-1275385, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097644} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.747057] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 729.747255] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 729.747430] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.747601] env[61824]: INFO nova.compute.manager [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Took 1.08 seconds to destroy the instance on the hypervisor. [ 729.747841] env[61824]: DEBUG oslo.service.loopingcall [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.748045] env[61824]: DEBUG nova.compute.manager [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.748367] env[61824]: DEBUG nova.network.neutron [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 729.803030] env[61824]: DEBUG nova.compute.utils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.808381] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 729.808381] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.858160] env[61824]: DEBUG nova.policy [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd05e42bfa416428d81c19861b11eea2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '858b5ed95401468fbf97c4000f7302a5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.878513] env[61824]: DEBUG nova.network.neutron [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.071672] env[61824]: INFO nova.scheduler.client.report [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Deleted allocations for instance 662e815c-2548-4aed-a928-d98f9e1297e7 [ 730.306840] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 730.383182] env[61824]: DEBUG nova.network.neutron [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.393548] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Successfully created port: a5daeb9a-b1e3-486f-bda4-69bac545c088 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.582460] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0888d69f-98f7-46aa-8a30-3bdbf4dd1371 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "662e815c-2548-4aed-a928-d98f9e1297e7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.453s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.678679] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c562290-2370-4cda-bf06-7793507a2a83 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.687424] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0934f286-d403-4009-a92d-e04f1a4008fb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.723577] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264682ac-359c-4642-a4ba-94afff33a0d8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.732129] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b78d7b5-9680-4ad1-a85e-1bd70beb487b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.751034] env[61824]: DEBUG nova.compute.provider_tree [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.886758] env[61824]: INFO nova.compute.manager [-] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Took 1.14 seconds to deallocate network for instance. [ 731.084614] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 731.257189] env[61824]: DEBUG nova.scheduler.client.report [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.319605] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 731.330732] env[61824]: DEBUG nova.compute.manager [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Received event network-changed-a5daeb9a-b1e3-486f-bda4-69bac545c088 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.330732] env[61824]: DEBUG nova.compute.manager [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Refreshing instance network info cache due to event network-changed-a5daeb9a-b1e3-486f-bda4-69bac545c088. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 731.330732] env[61824]: DEBUG oslo_concurrency.lockutils [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] Acquiring lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.330732] env[61824]: DEBUG oslo_concurrency.lockutils [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] Acquired lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.331248] env[61824]: DEBUG nova.network.neutron [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Refreshing network info cache for port a5daeb9a-b1e3-486f-bda4-69bac545c088 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 731.353329] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 731.353574] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 731.353726] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.353903] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 731.354062] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.354212] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 731.354423] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 731.354580] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 731.354822] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 731.355024] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 731.355202] env[61824]: DEBUG nova.virt.hardware [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 731.356073] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c973315-c679-4a94-ab20-8fa67d166879 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.365169] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661aa239-cd8e-4c52-9a5b-911c2a5a3093 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.397187] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.503567] env[61824]: ERROR nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. [ 731.503567] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 731.503567] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.503567] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 731.503567] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.503567] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 731.503567] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.503567] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 731.503567] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.503567] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 731.503567] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.503567] env[61824]: ERROR nova.compute.manager raise self.value [ 731.503567] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.503567] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 731.503567] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.503567] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 731.504053] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.504053] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 731.504053] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. [ 731.504053] env[61824]: ERROR nova.compute.manager [ 731.504053] env[61824]: Traceback (most recent call last): [ 731.504053] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 731.504053] env[61824]: listener.cb(fileno) [ 731.504053] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.504053] env[61824]: result = function(*args, **kwargs) [ 731.504053] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.504053] env[61824]: return func(*args, **kwargs) [ 731.504053] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.504053] env[61824]: raise e [ 731.504053] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.504053] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 731.504053] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.504053] env[61824]: created_port_ids = self._update_ports_for_instance( [ 731.504053] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.504053] env[61824]: with excutils.save_and_reraise_exception(): [ 731.504053] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.504053] env[61824]: self.force_reraise() [ 731.504053] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.504053] env[61824]: raise self.value [ 731.504053] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.504053] env[61824]: updated_port = self._update_port( [ 731.504053] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.504053] env[61824]: _ensure_no_port_binding_failure(port) [ 731.504053] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.504053] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 731.504944] env[61824]: nova.exception.PortBindingFailed: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. [ 731.504944] env[61824]: Removing descriptor: 19 [ 731.504944] env[61824]: ERROR nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Traceback (most recent call last): [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] yield resources [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self.driver.spawn(context, instance, image_meta, [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.504944] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] vm_ref = self.build_virtual_machine(instance, [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] for vif in network_info: [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return self._sync_wrapper(fn, *args, **kwargs) [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self.wait() [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self[:] = self._gt.wait() [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return self._exit_event.wait() [ 731.505359] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] result = hub.switch() [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return self.greenlet.switch() [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] result = function(*args, **kwargs) [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return func(*args, **kwargs) [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] raise e [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] nwinfo = self.network_api.allocate_for_instance( [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 731.505821] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] created_port_ids = self._update_ports_for_instance( [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] with excutils.save_and_reraise_exception(): [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self.force_reraise() [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] raise self.value [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] updated_port = self._update_port( [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] _ensure_no_port_binding_failure(port) [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.506241] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] raise exception.PortBindingFailed(port_id=port['id']) [ 731.506616] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] nova.exception.PortBindingFailed: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. [ 731.506616] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] [ 731.506616] env[61824]: INFO nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Terminating instance [ 731.506947] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.607251] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.760498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.761013] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 731.763659] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.816s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.765469] env[61824]: INFO nova.compute.claims [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.850696] env[61824]: DEBUG nova.network.neutron [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.925206] env[61824]: DEBUG nova.network.neutron [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.270504] env[61824]: DEBUG nova.compute.utils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.271921] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 732.274204] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 732.327480] env[61824]: DEBUG nova.policy [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06d0a8c525d546aa99e6da484071265e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a63dc38b308f4c538d9a9e9237e38e30', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 732.427946] env[61824]: DEBUG oslo_concurrency.lockutils [req-9ee9ecbe-00d5-4954-a7a8-08a98f2979c5 req-70d3429b-b718-4079-8c35-c80c8720786a service nova] Releasing lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.429441] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquired lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.429441] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.657855] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Successfully created port: 929eaf52-3d9f-49d5-9408-3d6147894d02 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.782337] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.950386] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.033522] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.172071] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76160390-16d4-4385-960f-0b51448c4410 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.184016] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1ca5f9-c475-49d9-8cac-0ca9fff47dcc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.213685] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d9b1a0-a2f6-4128-a0b4-98b3b3fd08a8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.221926] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035cab26-ddf5-4b49-a641-7b421692d7b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.235755] env[61824]: DEBUG nova.compute.provider_tree [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.384148] env[61824]: DEBUG nova.compute.manager [req-72b7f8c5-532b-4dee-bc5e-7fa0c211ac9b req-02c9ffbc-b59e-4c46-8680-78c4c91e8d9a service nova] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Received event network-vif-deleted-a5daeb9a-b1e3-486f-bda4-69bac545c088 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.536025] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Releasing lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.536025] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 733.536025] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.536025] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e9c1c6a-c4c4-41ee-a819-36ef21ac97c5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.545116] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fc6340-e7f5-45d9-a905-4f1ec61271c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.570848] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9942dd4b-8c04-40ac-b9f4-318b13902535 could not be found. [ 733.571095] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.571336] env[61824]: INFO nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.571538] env[61824]: DEBUG oslo.service.loopingcall [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.571761] env[61824]: DEBUG nova.compute.manager [-] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.571855] env[61824]: DEBUG nova.network.neutron [-] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 733.597431] env[61824]: DEBUG nova.network.neutron [-] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.738753] env[61824]: DEBUG nova.scheduler.client.report [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.799015] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 733.827492] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.827733] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.827910] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.828072] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.828218] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.828360] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.828566] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.828890] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.828890] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.829082] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.829196] env[61824]: DEBUG nova.virt.hardware [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.830103] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b205da90-50b7-4907-b811-d8c9127ed636 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.838469] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f19d5d-edf1-4589-adf1-1f98f11debab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.858927] env[61824]: ERROR nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. [ 733.858927] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.858927] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.858927] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.858927] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.858927] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.858927] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.858927] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.858927] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.858927] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 733.858927] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.858927] env[61824]: ERROR nova.compute.manager raise self.value [ 733.858927] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.858927] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.858927] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.858927] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.859574] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.859574] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.859574] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. [ 733.859574] env[61824]: ERROR nova.compute.manager [ 733.859574] env[61824]: Traceback (most recent call last): [ 733.859574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.859574] env[61824]: listener.cb(fileno) [ 733.859574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.859574] env[61824]: result = function(*args, **kwargs) [ 733.859574] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.859574] env[61824]: return func(*args, **kwargs) [ 733.859574] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.859574] env[61824]: raise e [ 733.859574] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.859574] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 733.859574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.859574] env[61824]: created_port_ids = self._update_ports_for_instance( [ 733.859574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.859574] env[61824]: with excutils.save_and_reraise_exception(): [ 733.859574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.859574] env[61824]: self.force_reraise() [ 733.859574] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.859574] env[61824]: raise self.value [ 733.859574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.859574] env[61824]: updated_port = self._update_port( [ 733.859574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.859574] env[61824]: _ensure_no_port_binding_failure(port) [ 733.859574] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.859574] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.860505] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. [ 733.860505] env[61824]: Removing descriptor: 19 [ 733.860505] env[61824]: ERROR nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] Traceback (most recent call last): [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] yield resources [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self.driver.spawn(context, instance, image_meta, [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.860505] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] vm_ref = self.build_virtual_machine(instance, [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] for vif in network_info: [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return self._sync_wrapper(fn, *args, **kwargs) [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self.wait() [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self[:] = self._gt.wait() [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return self._exit_event.wait() [ 733.860892] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] result = hub.switch() [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return self.greenlet.switch() [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] result = function(*args, **kwargs) [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return func(*args, **kwargs) [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] raise e [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] nwinfo = self.network_api.allocate_for_instance( [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 733.861344] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] created_port_ids = self._update_ports_for_instance( [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] with excutils.save_and_reraise_exception(): [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self.force_reraise() [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] raise self.value [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] updated_port = self._update_port( [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] _ensure_no_port_binding_failure(port) [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.862179] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] raise exception.PortBindingFailed(port_id=port['id']) [ 733.862598] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] nova.exception.PortBindingFailed: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. [ 733.862598] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] [ 733.862598] env[61824]: INFO nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Terminating instance [ 733.862598] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Acquiring lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.862598] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Acquired lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.862598] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 734.100285] env[61824]: DEBUG nova.network.neutron [-] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.246964] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.247519] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 734.250328] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.208s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.251689] env[61824]: INFO nova.compute.claims [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 734.381555] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.455199] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.488704] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.488936] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.602397] env[61824]: INFO nova.compute.manager [-] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Took 1.03 seconds to deallocate network for instance. [ 734.604607] env[61824]: DEBUG nova.compute.claims [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.604774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.755636] env[61824]: DEBUG nova.compute.utils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.759064] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 734.759064] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 734.812741] env[61824]: DEBUG nova.policy [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1d7f006ab244da48e1bfc38a5859932', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e23c7b193932478096b2fbf49a59fd7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.961016] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Releasing lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.961016] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.961016] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 734.961016] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e94bdf4-2dd4-4439-a864-a235d87efbc1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.968381] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73856d33-9697-4be7-aa6e-8d3a3966b522 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.991154] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d889d359-f767-4853-8092-070e7656c284 could not be found. [ 734.991154] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.991335] env[61824]: INFO nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Took 0.03 seconds to destroy the instance on the hypervisor. [ 734.991593] env[61824]: DEBUG oslo.service.loopingcall [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.991796] env[61824]: DEBUG nova.compute.manager [-] [instance: d889d359-f767-4853-8092-070e7656c284] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.991884] env[61824]: DEBUG nova.network.neutron [-] [instance: d889d359-f767-4853-8092-070e7656c284] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 735.022194] env[61824]: DEBUG nova.network.neutron [-] [instance: d889d359-f767-4853-8092-070e7656c284] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.137711] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Successfully created port: d67303b7-c5f5-4b10-8bdd-83d63875492f {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.261887] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 735.524046] env[61824]: DEBUG nova.network.neutron [-] [instance: d889d359-f767-4853-8092-070e7656c284] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.544051] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0213fddd-2226-437c-bc83-9b529463f68c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.550136] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e087fbc-30a8-4048-b5f8-737c10eacc5a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.581717] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a8c572-3f72-4cfe-86ac-701f857eb1dd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.585126] env[61824]: DEBUG nova.compute.manager [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] [instance: d889d359-f767-4853-8092-070e7656c284] Received event network-changed-929eaf52-3d9f-49d5-9408-3d6147894d02 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.585331] env[61824]: DEBUG nova.compute.manager [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] [instance: d889d359-f767-4853-8092-070e7656c284] Refreshing instance network info cache due to event network-changed-929eaf52-3d9f-49d5-9408-3d6147894d02. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.585517] env[61824]: DEBUG oslo_concurrency.lockutils [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] Acquiring lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.585673] env[61824]: DEBUG oslo_concurrency.lockutils [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] Acquired lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.585812] env[61824]: DEBUG nova.network.neutron [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] [instance: d889d359-f767-4853-8092-070e7656c284] Refreshing network info cache for port 929eaf52-3d9f-49d5-9408-3d6147894d02 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 735.592682] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836c5fc3-62ca-4731-8e3e-2fc52866b210 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.608931] env[61824]: DEBUG nova.compute.provider_tree [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.972514] env[61824]: ERROR nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. [ 735.972514] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.972514] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.972514] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.972514] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.972514] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.972514] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.972514] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.972514] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.972514] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 735.972514] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.972514] env[61824]: ERROR nova.compute.manager raise self.value [ 735.972514] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.972514] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.972514] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.972514] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.973276] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.973276] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.973276] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. [ 735.973276] env[61824]: ERROR nova.compute.manager [ 735.973276] env[61824]: Traceback (most recent call last): [ 735.973276] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.973276] env[61824]: listener.cb(fileno) [ 735.973276] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.973276] env[61824]: result = function(*args, **kwargs) [ 735.973276] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 735.973276] env[61824]: return func(*args, **kwargs) [ 735.973276] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.973276] env[61824]: raise e [ 735.973276] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.973276] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 735.973276] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 735.973276] env[61824]: created_port_ids = self._update_ports_for_instance( [ 735.973276] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 735.973276] env[61824]: with excutils.save_and_reraise_exception(): [ 735.973276] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.973276] env[61824]: self.force_reraise() [ 735.973276] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.973276] env[61824]: raise self.value [ 735.973276] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 735.973276] env[61824]: updated_port = self._update_port( [ 735.973276] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.973276] env[61824]: _ensure_no_port_binding_failure(port) [ 735.973276] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.973276] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.974295] env[61824]: nova.exception.PortBindingFailed: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. [ 735.974295] env[61824]: Removing descriptor: 19 [ 736.026385] env[61824]: INFO nova.compute.manager [-] [instance: d889d359-f767-4853-8092-070e7656c284] Took 1.03 seconds to deallocate network for instance. [ 736.028946] env[61824]: DEBUG nova.compute.claims [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.029147] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.103499] env[61824]: DEBUG nova.network.neutron [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] [instance: d889d359-f767-4853-8092-070e7656c284] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.111021] env[61824]: DEBUG nova.scheduler.client.report [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.176407] env[61824]: DEBUG nova.network.neutron [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] [instance: d889d359-f767-4853-8092-070e7656c284] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.279096] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.304267] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.304516] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.304670] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.304851] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.304996] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.305158] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.305358] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.305511] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.305670] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.305829] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.306009] env[61824]: DEBUG nova.virt.hardware [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.306907] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40aa76ff-c1c7-4ca0-9e6f-3315bc9308b4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.314846] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796320d4-9e32-4928-b91e-7f9c82cf5ce3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.328589] env[61824]: ERROR nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Traceback (most recent call last): [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] yield resources [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self.driver.spawn(context, instance, image_meta, [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] vm_ref = self.build_virtual_machine(instance, [ 736.328589] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] for vif in network_info: [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] return self._sync_wrapper(fn, *args, **kwargs) [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self.wait() [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self[:] = self._gt.wait() [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] return self._exit_event.wait() [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.329057] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] current.throw(*self._exc) [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] result = function(*args, **kwargs) [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] return func(*args, **kwargs) [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] raise e [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] nwinfo = self.network_api.allocate_for_instance( [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] created_port_ids = self._update_ports_for_instance( [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] with excutils.save_and_reraise_exception(): [ 736.329572] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self.force_reraise() [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] raise self.value [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] updated_port = self._update_port( [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] _ensure_no_port_binding_failure(port) [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] raise exception.PortBindingFailed(port_id=port['id']) [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] nova.exception.PortBindingFailed: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. [ 736.330026] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] [ 736.330026] env[61824]: INFO nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Terminating instance [ 736.330872] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.331044] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.331228] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.615675] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.616234] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.619217] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.673s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.679795] env[61824]: DEBUG oslo_concurrency.lockutils [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] Releasing lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.680067] env[61824]: DEBUG nova.compute.manager [req-40d9ce40-9cba-4ae2-b3a6-0c50cebaf6bc req-52b64efc-f2a3-4abd-808f-bbe8755cfc05 service nova] [instance: d889d359-f767-4853-8092-070e7656c284] Received event network-vif-deleted-929eaf52-3d9f-49d5-9408-3d6147894d02 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.848442] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.933228] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.126960] env[61824]: DEBUG nova.compute.utils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.128894] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 737.129073] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 737.168451] env[61824]: DEBUG nova.policy [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b37608f29dd14ba5b8afdf8264b297b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48f5283ba1b5404a826e7d2c526eeabe', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.389596] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb94f370-cac9-4026-adaa-05a9405b36f0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.399150] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4c016a-f177-4114-854f-ec81f74374ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.430660] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Successfully created port: 5e83b251-7541-475e-ae4e-f62a26762f1d {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.432810] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0ad7e0-47d5-4346-9450-abad0eec4fef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.442020] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.442020] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 737.442020] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 737.442020] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2434bd8-483d-499b-bce0-bb7e3772a019 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.442531] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894f4bf2-fcb7-467e-aa9a-627125540573 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.455727] env[61824]: DEBUG nova.compute.provider_tree [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.461264] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13977b99-f2db-44bd-80ac-fac945157b62 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.482336] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61d1f216-7ff1-419d-b415-3dce3537b1b4 could not be found. [ 737.482562] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 737.482739] env[61824]: INFO nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 737.482977] env[61824]: DEBUG oslo.service.loopingcall [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 737.483202] env[61824]: DEBUG nova.compute.manager [-] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.483293] env[61824]: DEBUG nova.network.neutron [-] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 737.497843] env[61824]: DEBUG nova.network.neutron [-] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.607839] env[61824]: DEBUG nova.compute.manager [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Received event network-changed-d67303b7-c5f5-4b10-8bdd-83d63875492f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.608108] env[61824]: DEBUG nova.compute.manager [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Refreshing instance network info cache due to event network-changed-d67303b7-c5f5-4b10-8bdd-83d63875492f. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.608344] env[61824]: DEBUG oslo_concurrency.lockutils [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] Acquiring lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.608494] env[61824]: DEBUG oslo_concurrency.lockutils [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] Acquired lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.608702] env[61824]: DEBUG nova.network.neutron [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Refreshing network info cache for port d67303b7-c5f5-4b10-8bdd-83d63875492f {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.632635] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.959436] env[61824]: DEBUG nova.scheduler.client.report [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.002893] env[61824]: DEBUG nova.network.neutron [-] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.129011] env[61824]: DEBUG nova.network.neutron [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.218246] env[61824]: DEBUG nova.network.neutron [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.230364] env[61824]: ERROR nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. [ 738.230364] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.230364] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.230364] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.230364] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.230364] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.230364] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.230364] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.230364] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.230364] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 738.230364] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.230364] env[61824]: ERROR nova.compute.manager raise self.value [ 738.230364] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.230364] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.230364] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.230364] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.231024] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.231024] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.231024] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. [ 738.231024] env[61824]: ERROR nova.compute.manager [ 738.231024] env[61824]: Traceback (most recent call last): [ 738.231024] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.231024] env[61824]: listener.cb(fileno) [ 738.231024] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.231024] env[61824]: result = function(*args, **kwargs) [ 738.231024] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.231024] env[61824]: return func(*args, **kwargs) [ 738.231024] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.231024] env[61824]: raise e [ 738.231024] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.231024] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 738.231024] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.231024] env[61824]: created_port_ids = self._update_ports_for_instance( [ 738.231024] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.231024] env[61824]: with excutils.save_and_reraise_exception(): [ 738.231024] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.231024] env[61824]: self.force_reraise() [ 738.231024] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.231024] env[61824]: raise self.value [ 738.231024] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.231024] env[61824]: updated_port = self._update_port( [ 738.231024] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.231024] env[61824]: _ensure_no_port_binding_failure(port) [ 738.231024] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.231024] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.232236] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. [ 738.232236] env[61824]: Removing descriptor: 19 [ 738.467623] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.848s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.468290] env[61824]: ERROR nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Traceback (most recent call last): [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self.driver.spawn(context, instance, image_meta, [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] vm_ref = self.build_virtual_machine(instance, [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.468290] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] for vif in network_info: [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] return self._sync_wrapper(fn, *args, **kwargs) [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self.wait() [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self[:] = self._gt.wait() [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] return self._exit_event.wait() [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] current.throw(*self._exc) [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.468713] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] result = function(*args, **kwargs) [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] return func(*args, **kwargs) [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] raise e [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] nwinfo = self.network_api.allocate_for_instance( [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] created_port_ids = self._update_ports_for_instance( [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] with excutils.save_and_reraise_exception(): [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] self.force_reraise() [ 738.469147] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] raise self.value [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] updated_port = self._update_port( [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] _ensure_no_port_binding_failure(port) [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] raise exception.PortBindingFailed(port_id=port['id']) [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] nova.exception.PortBindingFailed: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. [ 738.469572] env[61824]: ERROR nova.compute.manager [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] [ 738.469572] env[61824]: DEBUG nova.compute.utils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.470242] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.591s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.473429] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Build of instance 204fd4be-4a16-4838-a30d-c78339b3834d was re-scheduled: Binding failed for port 14a7d012-82d8-4fed-a618-cd3336d5b57e, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.473656] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.473904] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquiring lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.474064] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Acquired lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.474224] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.505294] env[61824]: INFO nova.compute.manager [-] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Took 1.02 seconds to deallocate network for instance. [ 738.507676] env[61824]: DEBUG nova.compute.claims [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 738.507849] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.646149] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.669632] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.669857] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.670014] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.670200] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.670342] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.670485] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.670682] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.670836] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.670996] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.671169] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.671335] env[61824]: DEBUG nova.virt.hardware [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.672178] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73a1690-26fc-4256-8cfd-e8fb956f77eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.680057] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae124fc5-fcb8-4195-8482-a405717115b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.693074] env[61824]: ERROR nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Traceback (most recent call last): [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] yield resources [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self.driver.spawn(context, instance, image_meta, [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] vm_ref = self.build_virtual_machine(instance, [ 738.693074] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] for vif in network_info: [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] return self._sync_wrapper(fn, *args, **kwargs) [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self.wait() [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self[:] = self._gt.wait() [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] return self._exit_event.wait() [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 738.693490] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] current.throw(*self._exc) [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] result = function(*args, **kwargs) [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] return func(*args, **kwargs) [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] raise e [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] nwinfo = self.network_api.allocate_for_instance( [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] created_port_ids = self._update_ports_for_instance( [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] with excutils.save_and_reraise_exception(): [ 738.693929] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self.force_reraise() [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] raise self.value [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] updated_port = self._update_port( [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] _ensure_no_port_binding_failure(port) [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] raise exception.PortBindingFailed(port_id=port['id']) [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] nova.exception.PortBindingFailed: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. [ 738.694366] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] [ 738.694366] env[61824]: INFO nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Terminating instance [ 738.695551] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Acquiring lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.695703] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Acquired lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.695865] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.721743] env[61824]: DEBUG oslo_concurrency.lockutils [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] Releasing lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.721974] env[61824]: DEBUG nova.compute.manager [req-156fd136-49e5-4870-ab6b-42e9aad1599c req-d2fe5aa4-35c8-4864-b8f1-ec0fdc6b15d4 service nova] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Received event network-vif-deleted-d67303b7-c5f5-4b10-8bdd-83d63875492f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.994101] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.077226] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.213436] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.236817] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a080add-16ae-44b9-bc85-b380525dc7ac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.244318] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8d4d73-07b8-44a1-aac6-6b267eb0a256 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.276647] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9423da93-fc57-48cf-a16b-549901c1b909 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.283761] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bd8fc6-7ef9-4d85-bdf3-1b09eb67c39e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.296733] env[61824]: DEBUG nova.compute.provider_tree [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.312983] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.580081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Releasing lock "refresh_cache-204fd4be-4a16-4838-a30d-c78339b3834d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.580345] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.580526] env[61824]: DEBUG nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.580691] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.596055] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.631681] env[61824]: DEBUG nova.compute.manager [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Received event network-changed-5e83b251-7541-475e-ae4e-f62a26762f1d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.631920] env[61824]: DEBUG nova.compute.manager [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Refreshing instance network info cache due to event network-changed-5e83b251-7541-475e-ae4e-f62a26762f1d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.632088] env[61824]: DEBUG oslo_concurrency.lockutils [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] Acquiring lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.799692] env[61824]: DEBUG nova.scheduler.client.report [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.814975] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Releasing lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.815387] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.815571] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.815862] env[61824]: DEBUG oslo_concurrency.lockutils [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] Acquired lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.816040] env[61824]: DEBUG nova.network.neutron [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Refreshing network info cache for port 5e83b251-7541-475e-ae4e-f62a26762f1d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 739.817297] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c9f523b0-4ecf-4a92-b7ec-df91aec53189 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.826890] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fa895c-db04-4c98-af92-3a65e92a4325 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.848010] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 329b3d0d-7007-40e3-a554-fbd0c7497b69 could not be found. [ 739.848228] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.848403] env[61824]: INFO nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Took 0.03 seconds to destroy the instance on the hypervisor. [ 739.848633] env[61824]: DEBUG oslo.service.loopingcall [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.848828] env[61824]: DEBUG nova.compute.manager [-] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.848927] env[61824]: DEBUG nova.network.neutron [-] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.864268] env[61824]: DEBUG nova.network.neutron [-] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.098710] env[61824]: DEBUG nova.network.neutron [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.305019] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.835s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.305681] env[61824]: ERROR nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] Traceback (most recent call last): [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self.driver.spawn(context, instance, image_meta, [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] vm_ref = self.build_virtual_machine(instance, [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.305681] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] for vif in network_info: [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return self._sync_wrapper(fn, *args, **kwargs) [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self.wait() [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self[:] = self._gt.wait() [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return self._exit_event.wait() [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] result = hub.switch() [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.306088] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return self.greenlet.switch() [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] result = function(*args, **kwargs) [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] return func(*args, **kwargs) [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] raise e [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] nwinfo = self.network_api.allocate_for_instance( [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] created_port_ids = self._update_ports_for_instance( [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] with excutils.save_and_reraise_exception(): [ 740.306500] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] self.force_reraise() [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] raise self.value [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] updated_port = self._update_port( [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] _ensure_no_port_binding_failure(port) [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] raise exception.PortBindingFailed(port_id=port['id']) [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] nova.exception.PortBindingFailed: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. [ 740.307036] env[61824]: ERROR nova.compute.manager [instance: f525feb8-d24a-4f89-b640-471163d181ef] [ 740.307404] env[61824]: DEBUG nova.compute.utils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.307834] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.462s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.308058] env[61824]: DEBUG nova.objects.instance [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 740.311297] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Build of instance f525feb8-d24a-4f89-b640-471163d181ef was re-scheduled: Binding failed for port 256f6905-aca8-42b9-8586-32db2f6d3f5b, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 740.311771] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 740.312104] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.312326] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquired lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.312582] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 740.336355] env[61824]: DEBUG nova.network.neutron [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.370054] env[61824]: DEBUG nova.network.neutron [-] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.408167] env[61824]: DEBUG nova.network.neutron [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.601264] env[61824]: INFO nova.compute.manager [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] [instance: 204fd4be-4a16-4838-a30d-c78339b3834d] Took 1.02 seconds to deallocate network for instance. [ 740.834851] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.873379] env[61824]: INFO nova.compute.manager [-] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Took 1.02 seconds to deallocate network for instance. [ 740.875660] env[61824]: DEBUG nova.compute.claims [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 740.875828] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.910903] env[61824]: DEBUG oslo_concurrency.lockutils [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] Releasing lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.911200] env[61824]: DEBUG nova.compute.manager [req-1bde18b0-525a-444d-8c5b-563b55cd394c req-ff831047-519a-4f71-bbbd-22fe1ccdfbc4 service nova] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Received event network-vif-deleted-5e83b251-7541-475e-ae4e-f62a26762f1d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.926984] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.321748] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2212effa-3d0d-4c13-8923-b6d0e3088942 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.322848] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.363s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.324402] env[61824]: INFO nova.compute.claims [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.429613] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Releasing lock "refresh_cache-f525feb8-d24a-4f89-b640-471163d181ef" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.429854] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 741.430086] env[61824]: DEBUG nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.430294] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.446015] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.628927] env[61824]: INFO nova.scheduler.client.report [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Deleted allocations for instance 204fd4be-4a16-4838-a30d-c78339b3834d [ 741.948877] env[61824]: DEBUG nova.network.neutron [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.137381] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b1b110f-65cf-4076-900b-b4a9f46199e9 tempest-ListServerFiltersTestJSON-1081037889 tempest-ListServerFiltersTestJSON-1081037889-project-member] Lock "204fd4be-4a16-4838-a30d-c78339b3834d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.960s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.451233] env[61824]: INFO nova.compute.manager [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: f525feb8-d24a-4f89-b640-471163d181ef] Took 1.02 seconds to deallocate network for instance. [ 742.564425] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2027ce5-7c43-4cf8-b344-9581f344121a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.572061] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad34e93-3520-435f-873d-6687f41b7ff5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.602058] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540992d6-7d64-4e04-a05f-e2a7a99f2d9e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.609414] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b635c0a3-0c77-4791-aa3b-2af84df3b904 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.622213] env[61824]: DEBUG nova.compute.provider_tree [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.640637] env[61824]: DEBUG nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.125474] env[61824]: DEBUG nova.scheduler.client.report [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.158575] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.477374] env[61824]: INFO nova.scheduler.client.report [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Deleted allocations for instance f525feb8-d24a-4f89-b640-471163d181ef [ 743.630558] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.631191] env[61824]: DEBUG nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.634144] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.759s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.635548] env[61824]: INFO nova.compute.claims [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.990014] env[61824]: DEBUG oslo_concurrency.lockutils [None req-74203216-1709-4798-b531-50311879193d tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "f525feb8-d24a-4f89-b640-471163d181ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.595s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.139942] env[61824]: DEBUG nova.compute.utils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.143093] env[61824]: DEBUG nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Not allocating networking since 'none' was specified. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 744.492810] env[61824]: DEBUG nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 744.644050] env[61824]: DEBUG nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.910495] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9a01a6-691b-492a-8f95-50692a456cac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.917934] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a97223-eb32-4fa3-8cde-fbeebfde0e4e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.948802] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0edd323c-e6d5-4677-8eab-048304226d06 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.956059] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae41e62-0ac3-46bf-96a0-e9c0d51f64d7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.972245] env[61824]: DEBUG nova.compute.provider_tree [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.010291] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.476054] env[61824]: DEBUG nova.scheduler.client.report [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.560324] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "43071be0-eaab-415c-8dcb-948bd9702885" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.560606] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "43071be0-eaab-415c-8dcb-948bd9702885" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.658528] env[61824]: DEBUG nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.683113] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.683360] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.683516] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.683693] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.683834] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.683975] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.684195] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.684351] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.684511] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.684669] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.684835] env[61824]: DEBUG nova.virt.hardware [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.685738] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32452a9-891f-4e58-9215-e24f4920e8e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.694048] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67750a65-6a27-4bd8-9dca-9084d9ea779b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.706445] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 745.712029] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Creating folder: Project (0ed68bf2ff4a474b961859e244614c9e). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 745.712278] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d43846a5-4746-49ac-b4d6-20bce8749d96 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.721392] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Created folder: Project (0ed68bf2ff4a474b961859e244614c9e) in parent group-v274074. [ 745.721571] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Creating folder: Instances. Parent ref: group-v274098. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 745.721772] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15b878b2-c8ce-4ec3-ae1a-8efb1649e64c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.729618] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Created folder: Instances in parent group-v274098. [ 745.729832] env[61824]: DEBUG oslo.service.loopingcall [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.730016] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 745.730218] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b077f537-9a50-47df-9885-d9f1daf924e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.746088] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 745.746088] env[61824]: value = "task-1275388" [ 745.746088] env[61824]: _type = "Task" [ 745.746088] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.753083] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275388, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.980354] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.980828] env[61824]: DEBUG nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 745.983480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.952s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.983652] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.983797] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 745.984101] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.251s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.986018] env[61824]: INFO nova.compute.claims [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.990033] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9adc4305-832f-4f53-ac32-2ef88cc0aaa3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.997036] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf4345a-1828-49bc-ac53-8226bb629a68 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.010831] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac67842-4de1-4301-8bc8-c7735b2a5057 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.017652] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a35e1e-3e59-426d-ab78-c1a0b378336b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.048089] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181464MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 746.048261] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.256393] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275388, 'name': CreateVM_Task, 'duration_secs': 0.241396} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.256567] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 746.257048] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.257205] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.257520] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 746.257765] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13756775-ced4-4b57-870d-8d13c9dac2bd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.261808] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 746.261808] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]529f37a8-1ede-be10-d42d-b2610ccf92a3" [ 746.261808] env[61824]: _type = "Task" [ 746.261808] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.269227] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529f37a8-1ede-be10-d42d-b2610ccf92a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.490669] env[61824]: DEBUG nova.compute.utils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.492465] env[61824]: DEBUG nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Not allocating networking since 'none' was specified. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 746.772059] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529f37a8-1ede-be10-d42d-b2610ccf92a3, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.772385] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.772582] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.772806] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.772947] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.773134] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 746.773374] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6378aa9-11d9-44a1-9be2-f9bcfaa1f21b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.781036] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 746.781209] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 746.781883] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b632a626-eb2b-40fb-bd99-acf95ec621f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.786338] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 746.786338] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]522f762d-08f9-85b7-bf8d-579ccda48aef" [ 746.786338] env[61824]: _type = "Task" [ 746.786338] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.793385] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522f762d-08f9-85b7-bf8d-579ccda48aef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.995527] env[61824]: DEBUG nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.227515] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05f4365-138c-4b9d-a15e-e4619fdbbe0d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.234845] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba386700-4975-4bee-b46c-251e25241f97 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.264384] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b1515b-b68d-4c6f-8ff3-e567bc44ee74 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.271435] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca3393a-6d74-4af5-9cac-0ec67ea5c332 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.284627] env[61824]: DEBUG nova.compute.provider_tree [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.294658] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522f762d-08f9-85b7-bf8d-579ccda48aef, 'name': SearchDatastore_Task, 'duration_secs': 0.007791} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.295897] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2834551-df68-44b6-a0b4-5272a87dea18 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.301271] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 747.301271] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a13b8b-5d11-235e-208e-ed11b2d1a77f" [ 747.301271] env[61824]: _type = "Task" [ 747.301271] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.308898] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a13b8b-5d11-235e-208e-ed11b2d1a77f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.791578] env[61824]: DEBUG nova.scheduler.client.report [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 747.812218] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a13b8b-5d11-235e-208e-ed11b2d1a77f, 'name': SearchDatastore_Task, 'duration_secs': 0.009213} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.812481] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.812728] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 3f89fb2d-99c1-4cf7-83aa-db033211a2ce/3f89fb2d-99c1-4cf7-83aa-db033211a2ce.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 747.813029] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9cb9a9f4-0ce0-41d4-9b0b-1c65769302aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.819618] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 747.819618] env[61824]: value = "task-1275389" [ 747.819618] env[61824]: _type = "Task" [ 747.819618] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.827746] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275389, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.007944] env[61824]: DEBUG nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.031121] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.031383] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.031538] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.031735] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.031885] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.032048] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.032264] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.032424] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.032588] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.032747] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.032915] env[61824]: DEBUG nova.virt.hardware [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.033824] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2da133-c5ec-408b-a19a-0728706383ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.042480] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b18962-f735-4ad8-afbd-4cea31001ca6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.056559] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 748.063073] env[61824]: DEBUG oslo.service.loopingcall [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.063302] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 748.063523] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e4adacab-e8d6-4f7a-b645-b00584706535 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.080770] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 748.080770] env[61824]: value = "task-1275390" [ 748.080770] env[61824]: _type = "Task" [ 748.080770] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.088703] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275390, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.296192] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.296726] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.299576] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.903s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.299787] env[61824]: DEBUG nova.objects.instance [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lazy-loading 'resources' on Instance uuid e758c89f-e5e2-42e0-926e-c629cb496a89 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.330875] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275389, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45546} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.331196] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 3f89fb2d-99c1-4cf7-83aa-db033211a2ce/3f89fb2d-99c1-4cf7-83aa-db033211a2ce.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 748.331480] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.331750] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44daa536-abbf-4591-9fef-a801f70b8389 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.338137] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 748.338137] env[61824]: value = "task-1275391" [ 748.338137] env[61824]: _type = "Task" [ 748.338137] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.346611] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275391, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.591150] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275390, 'name': CreateVM_Task, 'duration_secs': 0.324683} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.591282] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 748.591694] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.591849] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.592173] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 748.592406] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e58ba116-f344-4ea4-81b8-31a6eaf1a6fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.596804] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 748.596804] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52670e9b-675d-bc78-d611-e33d5e6ef8d5" [ 748.596804] env[61824]: _type = "Task" [ 748.596804] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.604794] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52670e9b-675d-bc78-d611-e33d5e6ef8d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.803312] env[61824]: DEBUG nova.compute.utils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.807784] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 748.807906] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 748.850391] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275391, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06833} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.850651] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 748.851435] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd52c25-aabf-4189-b76a-100e9870422e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.872200] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 3f89fb2d-99c1-4cf7-83aa-db033211a2ce/3f89fb2d-99c1-4cf7-83aa-db033211a2ce.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 748.873614] env[61824]: DEBUG nova.policy [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '316c089785a44aa897040f3c0ef7a071', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0924bfa41b5e42d68818557d2b0ace7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 748.876887] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b303191-1298-4c0f-8fb8-c734b1e0a652 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.897348] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 748.897348] env[61824]: value = "task-1275392" [ 748.897348] env[61824]: _type = "Task" [ 748.897348] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.907551] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275392, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.078240] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f3efdc-c9c8-4eb4-a612-0743e32145b6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.086388] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d89f8a8-7882-4df3-bb90-d056d9d29eb5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.121562] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5b7bea-c483-4f3b-adc4-2a4f64e57325 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.132407] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9659ca-699d-461f-8901-9c882737d441 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.136195] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52670e9b-675d-bc78-d611-e33d5e6ef8d5, 'name': SearchDatastore_Task, 'duration_secs': 0.009565} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.136371] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.136698] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 749.136991] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.137122] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.137300] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 749.137861] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6854b417-44c5-4952-a041-c9fa59577d56 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.147194] env[61824]: DEBUG nova.compute.provider_tree [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.154542] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 749.154542] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 749.155191] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4a5da90-b2c5-4104-8188-1797d98186da {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.161671] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 749.161671] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1a4a6-6a9b-69e3-e6f2-04cdd75e0400" [ 749.161671] env[61824]: _type = "Task" [ 749.161671] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.174230] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1a4a6-6a9b-69e3-e6f2-04cdd75e0400, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.180313] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Successfully created port: 7be17eae-3159-4e87-b87f-6dbab37400f2 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.308603] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.408040] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275392, 'name': ReconfigVM_Task, 'duration_secs': 0.287225} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.408347] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 3f89fb2d-99c1-4cf7-83aa-db033211a2ce/3f89fb2d-99c1-4cf7-83aa-db033211a2ce.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 749.408989] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b266c72d-4510-49b6-a4c9-49f6bbd8a906 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.415448] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 749.415448] env[61824]: value = "task-1275393" [ 749.415448] env[61824]: _type = "Task" [ 749.415448] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.426064] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275393, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.650935] env[61824]: DEBUG nova.scheduler.client.report [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.672615] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1a4a6-6a9b-69e3-e6f2-04cdd75e0400, 'name': SearchDatastore_Task, 'duration_secs': 0.018278} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.673653] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f1d4726-6a4e-4482-a0ec-fd9ad27b1979 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.679748] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 749.679748] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f50aed-f68a-32f6-e1f7-14db53cece24" [ 749.679748] env[61824]: _type = "Task" [ 749.679748] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.687756] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f50aed-f68a-32f6-e1f7-14db53cece24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.928634] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275393, 'name': Rename_Task, 'duration_secs': 0.136149} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.930758] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 749.931107] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44ccdf9a-62ba-443c-adb6-b048cceaa5c7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.937834] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 749.937834] env[61824]: value = "task-1275394" [ 749.937834] env[61824]: _type = "Task" [ 749.937834] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.946400] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.996829] env[61824]: DEBUG nova.compute.manager [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Received event network-changed-7be17eae-3159-4e87-b87f-6dbab37400f2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.997083] env[61824]: DEBUG nova.compute.manager [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Refreshing instance network info cache due to event network-changed-7be17eae-3159-4e87-b87f-6dbab37400f2. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 749.997302] env[61824]: DEBUG oslo_concurrency.lockutils [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] Acquiring lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.997443] env[61824]: DEBUG oslo_concurrency.lockutils [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] Acquired lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.997601] env[61824]: DEBUG nova.network.neutron [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Refreshing network info cache for port 7be17eae-3159-4e87-b87f-6dbab37400f2 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 750.135481] env[61824]: ERROR nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. [ 750.135481] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.135481] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.135481] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.135481] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.135481] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.135481] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.135481] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.135481] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.135481] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 750.135481] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.135481] env[61824]: ERROR nova.compute.manager raise self.value [ 750.135481] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.135481] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.135481] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.135481] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.136027] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.136027] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.136027] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. [ 750.136027] env[61824]: ERROR nova.compute.manager [ 750.136027] env[61824]: Traceback (most recent call last): [ 750.136027] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.136027] env[61824]: listener.cb(fileno) [ 750.136027] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.136027] env[61824]: result = function(*args, **kwargs) [ 750.136027] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.136027] env[61824]: return func(*args, **kwargs) [ 750.136027] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.136027] env[61824]: raise e [ 750.136027] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.136027] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 750.136027] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.136027] env[61824]: created_port_ids = self._update_ports_for_instance( [ 750.136027] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.136027] env[61824]: with excutils.save_and_reraise_exception(): [ 750.136027] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.136027] env[61824]: self.force_reraise() [ 750.136027] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.136027] env[61824]: raise self.value [ 750.136027] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.136027] env[61824]: updated_port = self._update_port( [ 750.136027] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.136027] env[61824]: _ensure_no_port_binding_failure(port) [ 750.136027] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.136027] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.136987] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. [ 750.136987] env[61824]: Removing descriptor: 19 [ 750.156029] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.158622] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.551s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.160205] env[61824]: INFO nova.compute.claims [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.182283] env[61824]: INFO nova.scheduler.client.report [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Deleted allocations for instance e758c89f-e5e2-42e0-926e-c629cb496a89 [ 750.194300] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f50aed-f68a-32f6-e1f7-14db53cece24, 'name': SearchDatastore_Task, 'duration_secs': 0.008923} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.194928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.194928] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 750.195125] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9721c216-d542-4707-8042-121a341c1d88 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.201809] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 750.201809] env[61824]: value = "task-1275395" [ 750.201809] env[61824]: _type = "Task" [ 750.201809] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.209831] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275395, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.317899] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.343691] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.343979] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.344196] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.344389] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.344535] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.344681] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.344890] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.345226] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.345557] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.345881] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.346228] env[61824]: DEBUG nova.virt.hardware [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.347571] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0ee0659-f763-49af-8656-240059a316fe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.360896] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473f96df-0ea9-4971-8985-8ff8595eac07 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.377800] env[61824]: ERROR nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Traceback (most recent call last): [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] yield resources [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self.driver.spawn(context, instance, image_meta, [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] vm_ref = self.build_virtual_machine(instance, [ 750.377800] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] for vif in network_info: [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] return self._sync_wrapper(fn, *args, **kwargs) [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self.wait() [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self[:] = self._gt.wait() [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] return self._exit_event.wait() [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.378485] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] current.throw(*self._exc) [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] result = function(*args, **kwargs) [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] return func(*args, **kwargs) [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] raise e [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] nwinfo = self.network_api.allocate_for_instance( [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] created_port_ids = self._update_ports_for_instance( [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] with excutils.save_and_reraise_exception(): [ 750.378920] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self.force_reraise() [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] raise self.value [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] updated_port = self._update_port( [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] _ensure_no_port_binding_failure(port) [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] raise exception.PortBindingFailed(port_id=port['id']) [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] nova.exception.PortBindingFailed: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. [ 750.379344] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] [ 750.379344] env[61824]: INFO nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Terminating instance [ 750.380891] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.448739] env[61824]: DEBUG oslo_vmware.api [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275394, 'name': PowerOnVM_Task, 'duration_secs': 0.442185} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.449063] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 750.449283] env[61824]: INFO nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Took 4.79 seconds to spawn the instance on the hypervisor. [ 750.449460] env[61824]: DEBUG nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 750.450339] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cb7c23-4882-4a95-86b6-f77b91c56e58 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.517356] env[61824]: DEBUG nova.network.neutron [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.611499] env[61824]: DEBUG nova.network.neutron [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.691018] env[61824]: DEBUG oslo_concurrency.lockutils [None req-983eef20-fa6c-456e-af36-7ac0b63984b7 tempest-ServerShowV257Test-1142829048 tempest-ServerShowV257Test-1142829048-project-member] Lock "e758c89f-e5e2-42e0-926e-c629cb496a89" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.257s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.711439] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275395, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439818} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.711691] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 750.711898] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.712147] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c50915f-5707-4b1c-8f91-e3792abae795 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.718816] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 750.718816] env[61824]: value = "task-1275396" [ 750.718816] env[61824]: _type = "Task" [ 750.718816] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.726287] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275396, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.969565] env[61824]: INFO nova.compute.manager [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Took 25.02 seconds to build instance. [ 751.114590] env[61824]: DEBUG oslo_concurrency.lockutils [req-009970cd-08ae-4df5-83aa-6c2ad9454462 req-200fd23a-8e4b-4fb1-a531-b823b00ff1d9 service nova] Releasing lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.114976] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.115175] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.228184] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.190753} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.230762] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.232743] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad057a1-5e21-4a8a-81f5-19add5edbe2b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.257194] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.260036] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e435c6f-54c7-4417-88eb-95426a8657b6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.279681] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 751.279681] env[61824]: value = "task-1275397" [ 751.279681] env[61824]: _type = "Task" [ 751.279681] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.289759] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275397, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.471870] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1bf67c5e-b5c2-4c56-b35a-ca55da28b585 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.440s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.476294] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f65125f-ee90-4eae-b862-b6d272c7ab13 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.483874] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa192e9-5700-4722-83f9-e392c3015b63 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.517776] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf657933-f8bc-4d41-ba80-c3671342fdaf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.525621] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7ff20a-7939-439f-8646-35eb82c7576f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.540065] env[61824]: DEBUG nova.compute.provider_tree [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.633676] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.716170] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.789383] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275397, 'name': ReconfigVM_Task, 'duration_secs': 0.272692} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.789699] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 751.790671] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d0ca2498-e3da-4f90-8e8f-788919da49c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.796921] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 751.796921] env[61824]: value = "task-1275398" [ 751.796921] env[61824]: _type = "Task" [ 751.796921] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.804461] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275398, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.974939] env[61824]: DEBUG nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.045915] env[61824]: DEBUG nova.scheduler.client.report [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.050612] env[61824]: DEBUG nova.compute.manager [req-7313cd7d-d74b-4e4c-a5d5-d0727117d0af req-87ec8e49-39a6-440e-91de-b64611cf3496 service nova] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Received event network-vif-deleted-7be17eae-3159-4e87-b87f-6dbab37400f2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.218920] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.219348] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.219553] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.219958] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b752c7cc-78dc-440b-97c4-5de160d6569b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.228662] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0be773b-f626-4724-a629-6d5fb457cc9f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.251069] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34f72773-5950-4710-b3e8-5e5c4877f227 could not be found. [ 752.251372] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.251560] env[61824]: INFO nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.251799] env[61824]: DEBUG oslo.service.loopingcall [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.252015] env[61824]: DEBUG nova.compute.manager [-] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.252110] env[61824]: DEBUG nova.network.neutron [-] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.267664] env[61824]: DEBUG nova.network.neutron [-] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.305556] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275398, 'name': Rename_Task, 'duration_secs': 0.132436} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.305806] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.306052] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ab65a6e-123c-43ff-bbcd-3d2b27a76c1a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.312218] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 752.312218] env[61824]: value = "task-1275399" [ 752.312218] env[61824]: _type = "Task" [ 752.312218] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.323516] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275399, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.496112] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.551783] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.552401] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 752.555082] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.950s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.769987] env[61824]: DEBUG nova.network.neutron [-] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.823570] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275399, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.060330] env[61824]: DEBUG nova.compute.utils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 753.064794] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 753.064997] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 753.106729] env[61824]: DEBUG nova.policy [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7038212763254338b3560d1c36c14468', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '19a44b94e3b04227be288b3515f9abf0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 753.272175] env[61824]: INFO nova.compute.manager [-] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Took 1.02 seconds to deallocate network for instance. [ 753.274584] env[61824]: DEBUG nova.compute.claims [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.274711] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.323175] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec98eb62-e83c-4543-b4dd-f2e8c92eeaa3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.328992] env[61824]: DEBUG oslo_vmware.api [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275399, 'name': PowerOnVM_Task, 'duration_secs': 0.589023} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.329655] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.329876] env[61824]: INFO nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Took 5.32 seconds to spawn the instance on the hypervisor. [ 753.330066] env[61824]: DEBUG nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.330834] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119a0c13-5b70-4650-8e88-3132ab159f13 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.336144] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32095666-3e2c-4879-b626-509a8715a392 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.369198] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d1cd0b-ed46-4d18-86ce-2a5fbb5ea615 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.377183] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6507fa6-259f-4300-8d93-07235295afc1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.392423] env[61824]: DEBUG nova.compute.provider_tree [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.417545] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Successfully created port: 7cabf642-da4b-4493-afc6-87b23fe52d99 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 753.566475] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 753.857191] env[61824]: INFO nova.compute.manager [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Took 26.00 seconds to build instance. [ 753.898024] env[61824]: DEBUG nova.scheduler.client.report [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.197256] env[61824]: DEBUG nova.compute.manager [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Received event network-changed-7cabf642-da4b-4493-afc6-87b23fe52d99 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.197539] env[61824]: DEBUG nova.compute.manager [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Refreshing instance network info cache due to event network-changed-7cabf642-da4b-4493-afc6-87b23fe52d99. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 754.198119] env[61824]: DEBUG oslo_concurrency.lockutils [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] Acquiring lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.198119] env[61824]: DEBUG oslo_concurrency.lockutils [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] Acquired lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.198119] env[61824]: DEBUG nova.network.neutron [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Refreshing network info cache for port 7cabf642-da4b-4493-afc6-87b23fe52d99 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 754.323028] env[61824]: ERROR nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. [ 754.323028] env[61824]: ERROR nova.compute.manager Traceback (most recent call last): [ 754.323028] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.323028] env[61824]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 754.323028] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 754.323028] env[61824]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 754.323028] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 754.323028] env[61824]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 754.323028] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.323028] env[61824]: ERROR nova.compute.manager self.force_reraise() [ 754.323028] env[61824]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.323028] env[61824]: ERROR nova.compute.manager raise self.value [ 754.323028] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 754.323028] env[61824]: ERROR nova.compute.manager updated_port = self._update_port( [ 754.323028] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.323028] env[61824]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 754.323558] env[61824]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.323558] env[61824]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 754.323558] env[61824]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. [ 754.323558] env[61824]: ERROR nova.compute.manager [ 754.323558] env[61824]: Traceback (most recent call last): [ 754.323558] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 754.323558] env[61824]: listener.cb(fileno) [ 754.323558] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.323558] env[61824]: result = function(*args, **kwargs) [ 754.323558] env[61824]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 754.323558] env[61824]: return func(*args, **kwargs) [ 754.323558] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.323558] env[61824]: raise e [ 754.323558] env[61824]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.323558] env[61824]: nwinfo = self.network_api.allocate_for_instance( [ 754.323558] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 754.323558] env[61824]: created_port_ids = self._update_ports_for_instance( [ 754.323558] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 754.323558] env[61824]: with excutils.save_and_reraise_exception(): [ 754.323558] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.323558] env[61824]: self.force_reraise() [ 754.323558] env[61824]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.323558] env[61824]: raise self.value [ 754.323558] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 754.323558] env[61824]: updated_port = self._update_port( [ 754.323558] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.323558] env[61824]: _ensure_no_port_binding_failure(port) [ 754.323558] env[61824]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.323558] env[61824]: raise exception.PortBindingFailed(port_id=port['id']) [ 754.324414] env[61824]: nova.exception.PortBindingFailed: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. [ 754.324414] env[61824]: Removing descriptor: 19 [ 754.358816] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e1ba76d3-081a-450b-9852-23e9697a28d3 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "9fd8e236-3556-4b95-952f-3c324b896e29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.940s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.401042] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.845s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.401233] env[61824]: ERROR nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Traceback (most recent call last): [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self.driver.spawn(context, instance, image_meta, [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] vm_ref = self.build_virtual_machine(instance, [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.401233] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] for vif in network_info: [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return self._sync_wrapper(fn, *args, **kwargs) [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self.wait() [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self[:] = self._gt.wait() [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return self._exit_event.wait() [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] result = hub.switch() [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 754.401621] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return self.greenlet.switch() [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] result = function(*args, **kwargs) [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] return func(*args, **kwargs) [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] raise e [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] nwinfo = self.network_api.allocate_for_instance( [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] created_port_ids = self._update_ports_for_instance( [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] with excutils.save_and_reraise_exception(): [ 754.402051] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] self.force_reraise() [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] raise self.value [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] updated_port = self._update_port( [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] _ensure_no_port_binding_failure(port) [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] raise exception.PortBindingFailed(port_id=port['id']) [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] nova.exception.PortBindingFailed: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. [ 754.402487] env[61824]: ERROR nova.compute.manager [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] [ 754.402852] env[61824]: DEBUG nova.compute.utils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 754.403374] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.374s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.407620] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Build of instance 9942dd4b-8c04-40ac-b9f4-318b13902535 was re-scheduled: Binding failed for port a5daeb9a-b1e3-486f-bda4-69bac545c088, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 754.408079] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 754.408320] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.408469] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquired lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.408629] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 754.490032] env[61824]: INFO nova.compute.manager [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Rebuilding instance [ 754.529652] env[61824]: DEBUG nova.compute.manager [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 754.530586] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195dc2ea-3910-43e7-8c25-a37a56ccbe69 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.576815] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 754.596685] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.596927] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.597148] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.597337] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.597481] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.597626] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.597828] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.597986] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.598168] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.598330] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.598499] env[61824]: DEBUG nova.virt.hardware [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.599348] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24200f3f-5e8a-4746-8614-1473c5a2ff8e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.607146] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a79043f-2dcb-44aa-b761-757775d8eecc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.621750] env[61824]: ERROR nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Traceback (most recent call last): [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] yield resources [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self.driver.spawn(context, instance, image_meta, [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] vm_ref = self.build_virtual_machine(instance, [ 754.621750] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] for vif in network_info: [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] return self._sync_wrapper(fn, *args, **kwargs) [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self.wait() [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self[:] = self._gt.wait() [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] return self._exit_event.wait() [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 754.622309] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] current.throw(*self._exc) [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] result = function(*args, **kwargs) [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] return func(*args, **kwargs) [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] raise e [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] nwinfo = self.network_api.allocate_for_instance( [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] created_port_ids = self._update_ports_for_instance( [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] with excutils.save_and_reraise_exception(): [ 754.622840] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self.force_reraise() [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] raise self.value [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] updated_port = self._update_port( [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] _ensure_no_port_binding_failure(port) [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] raise exception.PortBindingFailed(port_id=port['id']) [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] nova.exception.PortBindingFailed: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. [ 754.623453] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] [ 754.623453] env[61824]: INFO nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Terminating instance [ 754.624012] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Acquiring lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.715478] env[61824]: DEBUG nova.network.neutron [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.787876] env[61824]: DEBUG nova.network.neutron [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.861356] env[61824]: DEBUG nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 754.928157] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.007523] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.040850] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 755.041153] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b34aaffb-ec97-4e77-86ca-67c2e867d254 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.048825] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 755.048825] env[61824]: value = "task-1275400" [ 755.048825] env[61824]: _type = "Task" [ 755.048825] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.056964] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275400, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.157236] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322add58-3b81-4be7-a44c-85fd88da18bd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.165197] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3391add1-114b-4614-8d99-1c795e070801 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.198195] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fce5ff-d4b2-4d6e-a4cb-0654ae455f8c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.205886] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0a7bb5-69be-45fd-8353-c850439f069e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.220480] env[61824]: DEBUG nova.compute.provider_tree [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.290643] env[61824]: DEBUG oslo_concurrency.lockutils [req-674a3433-fcd8-45de-ab73-c60ded7e90df req-4e74bf58-98c0-45e1-970e-69313cbb0c09 service nova] Releasing lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.291060] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Acquired lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.291255] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 755.378261] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.509987] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Releasing lock "refresh_cache-9942dd4b-8c04-40ac-b9f4-318b13902535" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.510255] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 755.510447] env[61824]: DEBUG nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.510609] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 755.526015] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.558159] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275400, 'name': PowerOffVM_Task, 'duration_secs': 0.12901} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.558380] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 755.560030] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.560030] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50afa795-2945-4116-858f-959008dfdcd3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.565319] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 755.565537] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36f367cf-ea24-4f8d-9069-1afff9dda748 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.589024] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 755.589206] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 755.589383] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleting the datastore file [datastore2] 9fd8e236-3556-4b95-952f-3c324b896e29 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 755.589625] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-863a01e7-6169-4bbc-a43c-d7ea8e5eb500 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.595493] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 755.595493] env[61824]: value = "task-1275402" [ 755.595493] env[61824]: _type = "Task" [ 755.595493] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.603033] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275402, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.723242] env[61824]: DEBUG nova.scheduler.client.report [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.807794] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.872100] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.028692] env[61824]: DEBUG nova.network.neutron [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.104409] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275402, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102774} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.104602] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 756.104778] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 756.104945] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 756.230061] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.826s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.230061] env[61824]: ERROR nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. [ 756.230061] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] Traceback (most recent call last): [ 756.230061] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.230061] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self.driver.spawn(context, instance, image_meta, [ 756.230061] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 756.230061] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.230061] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.230061] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] vm_ref = self.build_virtual_machine(instance, [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] for vif in network_info: [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return self._sync_wrapper(fn, *args, **kwargs) [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self.wait() [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self[:] = self._gt.wait() [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return self._exit_event.wait() [ 756.230718] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] result = hub.switch() [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return self.greenlet.switch() [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] result = function(*args, **kwargs) [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] return func(*args, **kwargs) [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] raise e [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] nwinfo = self.network_api.allocate_for_instance( [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 756.231171] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] created_port_ids = self._update_ports_for_instance( [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] with excutils.save_and_reraise_exception(): [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] self.force_reraise() [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] raise self.value [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] updated_port = self._update_port( [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] _ensure_no_port_binding_failure(port) [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.231589] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] raise exception.PortBindingFailed(port_id=port['id']) [ 756.231925] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] nova.exception.PortBindingFailed: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. [ 756.231925] env[61824]: ERROR nova.compute.manager [instance: d889d359-f767-4853-8092-070e7656c284] [ 756.231925] env[61824]: DEBUG nova.compute.utils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 756.231925] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.724s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.235778] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Build of instance d889d359-f767-4853-8092-070e7656c284 was re-scheduled: Binding failed for port 929eaf52-3d9f-49d5-9408-3d6147894d02, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 756.236298] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 756.236524] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Acquiring lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.236668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Acquired lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.236826] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 756.246814] env[61824]: DEBUG nova.compute.manager [req-fb35f601-7a38-4dd3-bfee-cd3aae37c2e9 req-66468762-a26f-478f-b20c-88891b912a77 service nova] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Received event network-vif-deleted-7cabf642-da4b-4493-afc6-87b23fe52d99 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.374765] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Releasing lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.375352] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 756.375594] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 756.375925] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3dbad3b4-8c87-4292-b4cb-00fb096fc4f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.385935] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0712212b-e1ef-4975-bf03-8c82ffc41660 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.408270] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f86f95d-2c41-47b2-93f8-c7e6983c913a could not be found. [ 756.408486] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 756.408660] env[61824]: INFO nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 756.408896] env[61824]: DEBUG oslo.service.loopingcall [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 756.409121] env[61824]: DEBUG nova.compute.manager [-] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.409217] env[61824]: DEBUG nova.network.neutron [-] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 756.424065] env[61824]: DEBUG nova.network.neutron [-] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.531115] env[61824]: INFO nova.compute.manager [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: 9942dd4b-8c04-40ac-b9f4-318b13902535] Took 1.02 seconds to deallocate network for instance. [ 756.754994] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 756.827716] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.926142] env[61824]: DEBUG nova.network.neutron [-] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.957923] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30635e36-e5a3-4037-99cd-a55a5fe39d73 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.965256] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ed1c49-b3b1-4011-a0cc-3ef58b9e7d8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.994307] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fe3178-a28e-4b76-b0cc-d422ad7e5419 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.000888] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd768492-4ff3-4f43-90c3-4ec2449f1861 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.013571] env[61824]: DEBUG nova.compute.provider_tree [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 757.135255] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.135509] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.135663] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.135841] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.135981] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.136142] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.136377] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.136547] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.136707] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.136862] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.137066] env[61824]: DEBUG nova.virt.hardware [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.138117] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb1a1f7-108d-4402-a67a-0070ac1b3f64 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.146336] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6962be-dfac-440c-9491-d3d9f11d468d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.158896] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.164332] env[61824]: DEBUG oslo.service.loopingcall [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.164538] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 757.164721] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc9f953f-96ec-47b3-a3ca-5a2ef8fbf2e4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.182118] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.182118] env[61824]: value = "task-1275403" [ 757.182118] env[61824]: _type = "Task" [ 757.182118] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.192492] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275403, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.331043] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Releasing lock "refresh_cache-d889d359-f767-4853-8092-070e7656c284" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.331043] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 757.331043] env[61824]: DEBUG nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.331533] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 757.348348] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.430192] env[61824]: INFO nova.compute.manager [-] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Took 1.02 seconds to deallocate network for instance. [ 757.433861] env[61824]: DEBUG nova.compute.claims [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Aborting claim: {{(pid=61824) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 757.434047] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.516738] env[61824]: DEBUG nova.scheduler.client.report [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 757.559578] env[61824]: INFO nova.scheduler.client.report [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Deleted allocations for instance 9942dd4b-8c04-40ac-b9f4-318b13902535 [ 757.693095] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275403, 'name': CreateVM_Task, 'duration_secs': 0.245617} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.693277] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 757.693696] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.693850] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.694202] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 757.694520] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30b7801c-842a-43f4-8431-ac23417fa296 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.698991] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 757.698991] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c38956-36b6-88b6-bc41-b8bbea36fa04" [ 757.698991] env[61824]: _type = "Task" [ 757.698991] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.706899] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c38956-36b6-88b6-bc41-b8bbea36fa04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.851867] env[61824]: DEBUG nova.network.neutron [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.022196] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.790s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.022729] env[61824]: ERROR nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Traceback (most recent call last): [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self.driver.spawn(context, instance, image_meta, [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] vm_ref = self.build_virtual_machine(instance, [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.022729] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] for vif in network_info: [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] return self._sync_wrapper(fn, *args, **kwargs) [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self.wait() [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self[:] = self._gt.wait() [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] return self._exit_event.wait() [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] current.throw(*self._exc) [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.023279] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] result = function(*args, **kwargs) [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] return func(*args, **kwargs) [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] raise e [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] nwinfo = self.network_api.allocate_for_instance( [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] created_port_ids = self._update_ports_for_instance( [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] with excutils.save_and_reraise_exception(): [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] self.force_reraise() [ 758.023852] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] raise self.value [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] updated_port = self._update_port( [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] _ensure_no_port_binding_failure(port) [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] raise exception.PortBindingFailed(port_id=port['id']) [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] nova.exception.PortBindingFailed: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. [ 758.024382] env[61824]: ERROR nova.compute.manager [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] [ 758.024382] env[61824]: DEBUG nova.compute.utils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 758.024985] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.149s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.028278] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Build of instance 61d1f216-7ff1-419d-b415-3dce3537b1b4 was re-scheduled: Binding failed for port d67303b7-c5f5-4b10-8bdd-83d63875492f, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 758.028700] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 758.028936] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.029099] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.029261] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.067333] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fda39ebc-5805-46a3-9cea-4d1341d68742 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "9942dd4b-8c04-40ac-b9f4-318b13902535" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.010s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.209293] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c38956-36b6-88b6-bc41-b8bbea36fa04, 'name': SearchDatastore_Task, 'duration_secs': 0.012927} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.209522] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.209742] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.209961] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.210122] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.210290] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.210531] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06af4db3-ff71-43c8-84d9-953ecb248e7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.218191] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.218350] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.219090] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac90e1a4-fe3c-437a-bef5-f98db5bc4a3f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.223521] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 758.223521] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52826669-f393-aa60-204c-64c7b6f81c86" [ 758.223521] env[61824]: _type = "Task" [ 758.223521] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.231890] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52826669-f393-aa60-204c-64c7b6f81c86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.355747] env[61824]: INFO nova.compute.manager [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] [instance: d889d359-f767-4853-8092-070e7656c284] Took 1.02 seconds to deallocate network for instance. [ 758.550036] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.570109] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.634799] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.735802] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52826669-f393-aa60-204c-64c7b6f81c86, 'name': SearchDatastore_Task, 'duration_secs': 0.009938} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.736587] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba66ed51-9d3c-41a4-a2f2-8ac33f4d7798 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.743768] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 758.743768] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52b3341f-40af-e552-ed92-46f51829856a" [ 758.743768] env[61824]: _type = "Task" [ 758.743768] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.751230] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b3341f-40af-e552-ed92-46f51829856a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.775734] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2f7870-3c8b-4581-a89a-5552392cfbf1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.783771] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6eb2cd-e1ad-4645-8d01-bbf8f7ab26af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.813676] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f70880-52bd-4848-b08e-dacf4f5943c7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.820287] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2ab430-a86a-459f-bcdd-ecd738090cb0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.832743] env[61824]: DEBUG nova.compute.provider_tree [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.094338] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.137435] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "refresh_cache-61d1f216-7ff1-419d-b415-3dce3537b1b4" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.137677] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 759.137863] env[61824]: DEBUG nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 759.138062] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.151840] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.254054] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b3341f-40af-e552-ed92-46f51829856a, 'name': SearchDatastore_Task, 'duration_secs': 0.009648} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.254442] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.254818] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 759.255090] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12059519-e427-4d00-99d6-f849638bc790 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.261488] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 759.261488] env[61824]: value = "task-1275404" [ 759.261488] env[61824]: _type = "Task" [ 759.261488] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.268570] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.338392] env[61824]: DEBUG nova.scheduler.client.report [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.387443] env[61824]: INFO nova.scheduler.client.report [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Deleted allocations for instance d889d359-f767-4853-8092-070e7656c284 [ 759.654709] env[61824]: DEBUG nova.network.neutron [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.771024] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275404, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448801} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.771200] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 759.771411] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 759.771659] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b5ba057-03c4-4e2d-a151-0995adade8a8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.778534] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 759.778534] env[61824]: value = "task-1275405" [ 759.778534] env[61824]: _type = "Task" [ 759.778534] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.786284] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275405, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.843816] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.819s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.844475] env[61824]: ERROR nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Traceback (most recent call last): [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self.driver.spawn(context, instance, image_meta, [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] vm_ref = self.build_virtual_machine(instance, [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.844475] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] for vif in network_info: [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] return self._sync_wrapper(fn, *args, **kwargs) [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self.wait() [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self[:] = self._gt.wait() [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] return self._exit_event.wait() [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] current.throw(*self._exc) [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.844853] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] result = function(*args, **kwargs) [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] return func(*args, **kwargs) [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] raise e [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] nwinfo = self.network_api.allocate_for_instance( [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] created_port_ids = self._update_ports_for_instance( [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] with excutils.save_and_reraise_exception(): [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] self.force_reraise() [ 759.845257] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] raise self.value [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] updated_port = self._update_port( [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] _ensure_no_port_binding_failure(port) [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] raise exception.PortBindingFailed(port_id=port['id']) [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] nova.exception.PortBindingFailed: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. [ 759.846089] env[61824]: ERROR nova.compute.manager [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] [ 759.846089] env[61824]: DEBUG nova.compute.utils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 759.846634] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.688s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.848296] env[61824]: INFO nova.compute.claims [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.850896] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Build of instance 329b3d0d-7007-40e3-a554-fbd0c7497b69 was re-scheduled: Binding failed for port 5e83b251-7541-475e-ae4e-f62a26762f1d, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 759.851396] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 759.851633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Acquiring lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.851781] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Acquired lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.851940] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.894633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-be77340c-6c93-40a1-a269-793339289178 tempest-InstanceActionsTestJSON-846126394 tempest-InstanceActionsTestJSON-846126394-project-member] Lock "d889d359-f767-4853-8092-070e7656c284" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.532s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.157924] env[61824]: INFO nova.compute.manager [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 61d1f216-7ff1-419d-b415-3dce3537b1b4] Took 1.02 seconds to deallocate network for instance. [ 760.288198] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275405, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067714} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.288489] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 760.289264] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7154276-702f-45a6-b2ff-c33d500142ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.308775] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.309042] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3c427c8-f5f0-4688-84ce-72954dabd770 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.328458] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 760.328458] env[61824]: value = "task-1275406" [ 760.328458] env[61824]: _type = "Task" [ 760.328458] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.338253] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275406, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.375618] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.397563] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.519706] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.736877] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.737175] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.838844] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275406, 'name': ReconfigVM_Task, 'duration_secs': 0.272113} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.839146] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 9fd8e236-3556-4b95-952f-3c324b896e29/9fd8e236-3556-4b95-952f-3c324b896e29.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.839854] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01c53f97-d0d3-4162-a1f5-4951c674cc8f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.847399] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 760.847399] env[61824]: value = "task-1275407" [ 760.847399] env[61824]: _type = "Task" [ 760.847399] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.855822] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275407, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.917584] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.026607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Releasing lock "refresh_cache-329b3d0d-7007-40e3-a554-fbd0c7497b69" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.026607] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 761.026607] env[61824]: DEBUG nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.026607] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.056582] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.195722] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d32edd1-df9d-4e17-a1fb-f8805f2de3f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.199969] env[61824]: INFO nova.scheduler.client.report [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted allocations for instance 61d1f216-7ff1-419d-b415-3dce3537b1b4 [ 761.210863] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a854763-b8b2-4508-b81b-9c6f404c3324 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.244319] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28abb54a-0291-4a6d-97d2-0874ac6a453c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.252211] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d855e46-4ba4-43d4-b8b4-77aef369c026 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.268731] env[61824]: DEBUG nova.compute.provider_tree [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.356831] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275407, 'name': Rename_Task, 'duration_secs': 0.14057} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.357457] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.357766] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d2b7969-2b46-46c8-9480-8881884f6601 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.363737] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 761.363737] env[61824]: value = "task-1275408" [ 761.363737] env[61824]: _type = "Task" [ 761.363737] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.371383] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275408, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.561291] env[61824]: DEBUG nova.network.neutron [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.708256] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c52db9ee-ff8b-4b2a-a8fe-4388c5079a06 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "61d1f216-7ff1-419d-b415-3dce3537b1b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.634s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.771791] env[61824]: DEBUG nova.scheduler.client.report [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.874166] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275408, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.065380] env[61824]: INFO nova.compute.manager [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] [instance: 329b3d0d-7007-40e3-a554-fbd0c7497b69] Took 1.04 seconds to deallocate network for instance. [ 762.211783] env[61824]: DEBUG nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 762.277147] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.277685] env[61824]: DEBUG nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 762.281490] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.270s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.281666] env[61824]: INFO nova.compute.claims [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.374684] env[61824]: DEBUG oslo_vmware.api [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275408, 'name': PowerOnVM_Task, 'duration_secs': 0.716361} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.375013] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 762.375324] env[61824]: DEBUG nova.compute.manager [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.376128] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cfd6164-fa69-472f-81f2-a7f48a55063f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.733963] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.790859] env[61824]: DEBUG nova.compute.utils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.792373] env[61824]: DEBUG nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 762.792679] env[61824]: DEBUG nova.network.neutron [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 762.850732] env[61824]: DEBUG nova.policy [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff4fff56a7cc4f19b92786ea07b66e8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bad9de0b032c47429ba879d2d62c6af6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.891879] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.097242] env[61824]: INFO nova.scheduler.client.report [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Deleted allocations for instance 329b3d0d-7007-40e3-a554-fbd0c7497b69 [ 763.134643] env[61824]: DEBUG nova.network.neutron [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Successfully created port: 51fba888-670e-4f85-97a5-0133e0624342 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.296737] env[61824]: DEBUG nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 763.454603] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "2495fcb4-28cb-49c3-90fe-c84072466287" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.455549] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "2495fcb4-28cb-49c3-90fe-c84072466287" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.528472] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e576f4ec-7bce-4d7b-8140-fea8c0b74edb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.531614] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "9fd8e236-3556-4b95-952f-3c324b896e29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.531857] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "9fd8e236-3556-4b95-952f-3c324b896e29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.532134] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "9fd8e236-3556-4b95-952f-3c324b896e29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.532370] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "9fd8e236-3556-4b95-952f-3c324b896e29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.532598] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "9fd8e236-3556-4b95-952f-3c324b896e29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.534798] env[61824]: INFO nova.compute.manager [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Terminating instance [ 763.538325] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "refresh_cache-9fd8e236-3556-4b95-952f-3c324b896e29" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.538541] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "refresh_cache-9fd8e236-3556-4b95-952f-3c324b896e29" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.538757] env[61824]: DEBUG nova.network.neutron [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 763.540397] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd5cf65-b682-4854-bdfa-22fcc4256573 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.572337] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076b64a6-23c3-44cb-9ead-80a6f0031ead {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.579552] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-267f4bb6-7b51-4a40-8338-b9cf344efe7b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.593408] env[61824]: DEBUG nova.compute.provider_tree [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.605222] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6ca8c728-28af-4361-be25-6923dd7e274a tempest-ServerRescueTestJSONUnderV235-1760672907 tempest-ServerRescueTestJSONUnderV235-1760672907-project-member] Lock "329b3d0d-7007-40e3-a554-fbd0c7497b69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.540s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.058214] env[61824]: DEBUG nova.network.neutron [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.095828] env[61824]: DEBUG nova.scheduler.client.report [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.107261] env[61824]: DEBUG nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 764.128424] env[61824]: DEBUG nova.network.neutron [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.305459] env[61824]: DEBUG nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 764.335064] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.335318] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.336186] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.336186] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.336186] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.336186] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.336186] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.336405] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.336448] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.336585] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.336792] env[61824]: DEBUG nova.virt.hardware [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.337704] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd50d38-e675-4fa4-9677-ec70b5fb6610 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.348019] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dffd39c-f22d-408d-b85a-d6fb80721ebb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.603997] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.603997] env[61824]: DEBUG nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.604837] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.557s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.632788] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "refresh_cache-9fd8e236-3556-4b95-952f-3c324b896e29" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.634126] env[61824]: DEBUG nova.compute.manager [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 764.634453] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.639602] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.639602] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b7b84c0-4180-44d4-85d8-bdd513a728b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.650445] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 764.650686] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-09ad4ef3-fbcf-4a4d-a281-e6c16c3694a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.656444] env[61824]: DEBUG oslo_vmware.api [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 764.656444] env[61824]: value = "task-1275409" [ 764.656444] env[61824]: _type = "Task" [ 764.656444] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.664668] env[61824]: DEBUG oslo_vmware.api [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275409, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.667223] env[61824]: DEBUG nova.compute.manager [req-b0520e24-8d07-47a9-9505-9d4112104ea3 req-6869fd28-09f8-413a-869f-4c4195e04063 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Received event network-vif-plugged-51fba888-670e-4f85-97a5-0133e0624342 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.667425] env[61824]: DEBUG oslo_concurrency.lockutils [req-b0520e24-8d07-47a9-9505-9d4112104ea3 req-6869fd28-09f8-413a-869f-4c4195e04063 service nova] Acquiring lock "7e73f8b5-a138-4455-a392-9a2b2b860558-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.667617] env[61824]: DEBUG oslo_concurrency.lockutils [req-b0520e24-8d07-47a9-9505-9d4112104ea3 req-6869fd28-09f8-413a-869f-4c4195e04063 service nova] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.667774] env[61824]: DEBUG oslo_concurrency.lockutils [req-b0520e24-8d07-47a9-9505-9d4112104ea3 req-6869fd28-09f8-413a-869f-4c4195e04063 service nova] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.668023] env[61824]: DEBUG nova.compute.manager [req-b0520e24-8d07-47a9-9505-9d4112104ea3 req-6869fd28-09f8-413a-869f-4c4195e04063 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] No waiting events found dispatching network-vif-plugged-51fba888-670e-4f85-97a5-0133e0624342 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 764.668124] env[61824]: WARNING nova.compute.manager [req-b0520e24-8d07-47a9-9505-9d4112104ea3 req-6869fd28-09f8-413a-869f-4c4195e04063 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Received unexpected event network-vif-plugged-51fba888-670e-4f85-97a5-0133e0624342 for instance with vm_state building and task_state spawning. [ 764.761226] env[61824]: DEBUG nova.network.neutron [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Successfully updated port: 51fba888-670e-4f85-97a5-0133e0624342 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 765.111411] env[61824]: DEBUG nova.compute.utils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.117040] env[61824]: DEBUG nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.117040] env[61824]: DEBUG nova.network.neutron [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 765.168019] env[61824]: DEBUG oslo_vmware.api [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275409, 'name': PowerOffVM_Task, 'duration_secs': 0.154655} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.169197] env[61824]: DEBUG nova.policy [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff4fff56a7cc4f19b92786ea07b66e8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bad9de0b032c47429ba879d2d62c6af6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.171014] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 765.171262] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 765.171527] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96b7a26a-92bd-4843-a378-ad6cc70b9b2e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.198892] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 765.198892] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 765.198892] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleting the datastore file [datastore1] 9fd8e236-3556-4b95-952f-3c324b896e29 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.198892] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8eaf81d-4427-4c05-b0cc-cd6df6ad8a5e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.202705] env[61824]: DEBUG oslo_vmware.api [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 765.202705] env[61824]: value = "task-1275411" [ 765.202705] env[61824]: _type = "Task" [ 765.202705] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.215441] env[61824]: DEBUG oslo_vmware.api [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275411, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.264825] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.265043] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.265181] env[61824]: DEBUG nova.network.neutron [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 765.421805] env[61824]: DEBUG nova.network.neutron [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Successfully created port: 64ce8844-4f88-487b-8fba-e00413a3eb05 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.616898] env[61824]: DEBUG nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.642522] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 3f89fb2d-99c1-4cf7-83aa-db033211a2ce actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.642682] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 9fd8e236-3556-4b95-952f-3c324b896e29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.642849] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 34f72773-5950-4710-b3e8-5e5c4877f227 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.642977] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 0f86f95d-2c41-47b2-93f8-c7e6983c913a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.643116] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7e73f8b5-a138-4455-a392-9a2b2b860558 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.643235] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7f6a4446-86e1-44c7-ab42-297b033dbace actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 765.711620] env[61824]: DEBUG oslo_vmware.api [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275411, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093512} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.711864] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 765.712061] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 765.712244] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.712412] env[61824]: INFO nova.compute.manager [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Took 1.08 seconds to destroy the instance on the hypervisor. [ 765.712643] env[61824]: DEBUG oslo.service.loopingcall [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.712824] env[61824]: DEBUG nova.compute.manager [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.712914] env[61824]: DEBUG nova.network.neutron [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 765.731444] env[61824]: DEBUG nova.network.neutron [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.806556] env[61824]: DEBUG nova.network.neutron [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.993595] env[61824]: DEBUG nova.network.neutron [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Updating instance_info_cache with network_info: [{"id": "51fba888-670e-4f85-97a5-0133e0624342", "address": "fa:16:3e:d2:52:4a", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51fba888-67", "ovs_interfaceid": "51fba888-670e-4f85-97a5-0133e0624342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.146605] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 23fbfb1e-a716-4a73-a15d-16ff25690e0e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.234124] env[61824]: DEBUG nova.network.neutron [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.495846] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.496179] env[61824]: DEBUG nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance network_info: |[{"id": "51fba888-670e-4f85-97a5-0133e0624342", "address": "fa:16:3e:d2:52:4a", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51fba888-67", "ovs_interfaceid": "51fba888-670e-4f85-97a5-0133e0624342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 766.497020] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:52:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6076d24d-3c8e-4bbb-ba96-a08fb27a73cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51fba888-670e-4f85-97a5-0133e0624342', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 766.504808] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating folder: Project (bad9de0b032c47429ba879d2d62c6af6). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.505679] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45838202-7b12-4817-bd81-5a36195c2254 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.517885] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created folder: Project (bad9de0b032c47429ba879d2d62c6af6) in parent group-v274074. [ 766.518080] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating folder: Instances. Parent ref: group-v274103. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 766.518324] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-413495cd-17c7-45da-b5af-f1152d77c5da {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.525966] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created folder: Instances in parent group-v274103. [ 766.526199] env[61824]: DEBUG oslo.service.loopingcall [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.526380] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 766.526564] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38a3c786-cbf3-49d3-a480-fa667d212775 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.544880] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 766.544880] env[61824]: value = "task-1275414" [ 766.544880] env[61824]: _type = "Task" [ 766.544880] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.554915] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275414, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.629515] env[61824]: DEBUG nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.649546] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance cc13d6a5-6b99-40bf-8978-09d284b4ed17 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.653525] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.653751] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.653908] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.654106] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.654258] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.654404] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.654609] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.654765] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.654924] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.655095] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.655267] env[61824]: DEBUG nova.virt.hardware [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.656128] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d54cb3-a685-47ba-ad4f-7521cd241bb0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.664469] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a6e63f-7377-4a3e-9c6f-8f989819892c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.692577] env[61824]: DEBUG nova.compute.manager [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Received event network-changed-51fba888-670e-4f85-97a5-0133e0624342 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.692805] env[61824]: DEBUG nova.compute.manager [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Refreshing instance network info cache due to event network-changed-51fba888-670e-4f85-97a5-0133e0624342. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.693103] env[61824]: DEBUG oslo_concurrency.lockutils [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] Acquiring lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.693315] env[61824]: DEBUG oslo_concurrency.lockutils [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] Acquired lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.693504] env[61824]: DEBUG nova.network.neutron [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Refreshing network info cache for port 51fba888-670e-4f85-97a5-0133e0624342 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 766.736963] env[61824]: INFO nova.compute.manager [-] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Took 1.02 seconds to deallocate network for instance. [ 766.876428] env[61824]: DEBUG nova.network.neutron [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Successfully updated port: 64ce8844-4f88-487b-8fba-e00413a3eb05 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 767.055418] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275414, 'name': CreateVM_Task, 'duration_secs': 0.277255} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.055603] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 767.062692] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.062862] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.063207] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 767.063448] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0891b45a-5b9b-458f-8959-b65093a54dd6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.068046] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 767.068046] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]523c7257-9a94-fddd-4093-c31355e5726d" [ 767.068046] env[61824]: _type = "Task" [ 767.068046] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.075177] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523c7257-9a94-fddd-4093-c31355e5726d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.153046] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance f24881d3-a647-42f9-bf82-0065f3425a5f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.243327] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.379257] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.379435] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.379562] env[61824]: DEBUG nova.network.neutron [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.424362] env[61824]: DEBUG nova.network.neutron [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Updated VIF entry in instance network info cache for port 51fba888-670e-4f85-97a5-0133e0624342. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 767.424688] env[61824]: DEBUG nova.network.neutron [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Updating instance_info_cache with network_info: [{"id": "51fba888-670e-4f85-97a5-0133e0624342", "address": "fa:16:3e:d2:52:4a", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51fba888-67", "ovs_interfaceid": "51fba888-670e-4f85-97a5-0133e0624342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.578778] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523c7257-9a94-fddd-4093-c31355e5726d, 'name': SearchDatastore_Task, 'duration_secs': 0.008856} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.578778] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.578778] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 767.578976] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.578976] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.579199] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 767.579447] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e507821c-05c0-454e-a6cf-fc89e1a625ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.586790] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 767.586966] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 767.587647] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c63c16e8-d78f-4a54-8b2b-18e523511194 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.592296] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 767.592296] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]527ba926-ae5a-8540-b9b7-12fa1d165e8c" [ 767.592296] env[61824]: _type = "Task" [ 767.592296] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.599802] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527ba926-ae5a-8540-b9b7-12fa1d165e8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.655646] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance e3e19a04-a29b-4473-833c-d8486703e9ad has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.912189] env[61824]: DEBUG nova.network.neutron [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.928836] env[61824]: DEBUG oslo_concurrency.lockutils [req-42f12ce6-e1e8-4f96-86b3-558c58d48050 req-07d4c6f9-1838-4b38-bfc2-4800d3331354 service nova] Releasing lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.032260] env[61824]: DEBUG nova.network.neutron [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Updating instance_info_cache with network_info: [{"id": "64ce8844-4f88-487b-8fba-e00413a3eb05", "address": "fa:16:3e:05:3e:0f", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64ce8844-4f", "ovs_interfaceid": "64ce8844-4f88-487b-8fba-e00413a3eb05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.101827] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527ba926-ae5a-8540-b9b7-12fa1d165e8c, 'name': SearchDatastore_Task, 'duration_secs': 0.007725} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.102583] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9a87aa2-cd20-4b89-9e28-4e9db36b2968 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.107842] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 768.107842] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f3926a-53a7-52ee-793a-be7c3b20864e" [ 768.107842] env[61824]: _type = "Task" [ 768.107842] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.115740] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f3926a-53a7-52ee-793a-be7c3b20864e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.158728] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.535358] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.535671] env[61824]: DEBUG nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Instance network_info: |[{"id": "64ce8844-4f88-487b-8fba-e00413a3eb05", "address": "fa:16:3e:05:3e:0f", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64ce8844-4f", "ovs_interfaceid": "64ce8844-4f88-487b-8fba-e00413a3eb05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 768.536107] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:3e:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6076d24d-3c8e-4bbb-ba96-a08fb27a73cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '64ce8844-4f88-487b-8fba-e00413a3eb05', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 768.543640] env[61824]: DEBUG oslo.service.loopingcall [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.543849] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 768.544101] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e374f9dd-56a1-4867-a99c-19b2ea15ce08 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.564496] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 768.564496] env[61824]: value = "task-1275415" [ 768.564496] env[61824]: _type = "Task" [ 768.564496] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.571928] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275415, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.617406] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f3926a-53a7-52ee-793a-be7c3b20864e, 'name': SearchDatastore_Task, 'duration_secs': 0.009878} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.617682] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.617967] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 768.618212] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9346bb5-213f-463f-9fe2-0facf6769969 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.624912] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 768.624912] env[61824]: value = "task-1275416" [ 768.624912] env[61824]: _type = "Task" [ 768.624912] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.632260] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275416, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.662267] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8522fe80-4d6d-4474-ae63-00daa0dc7349 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.723589] env[61824]: DEBUG nova.compute.manager [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Received event network-vif-plugged-64ce8844-4f88-487b-8fba-e00413a3eb05 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.723888] env[61824]: DEBUG oslo_concurrency.lockutils [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] Acquiring lock "7f6a4446-86e1-44c7-ab42-297b033dbace-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.724336] env[61824]: DEBUG oslo_concurrency.lockutils [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.724576] env[61824]: DEBUG oslo_concurrency.lockutils [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.724832] env[61824]: DEBUG nova.compute.manager [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] No waiting events found dispatching network-vif-plugged-64ce8844-4f88-487b-8fba-e00413a3eb05 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 768.725065] env[61824]: WARNING nova.compute.manager [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Received unexpected event network-vif-plugged-64ce8844-4f88-487b-8fba-e00413a3eb05 for instance with vm_state building and task_state spawning. [ 768.725316] env[61824]: DEBUG nova.compute.manager [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Received event network-changed-64ce8844-4f88-487b-8fba-e00413a3eb05 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.725537] env[61824]: DEBUG nova.compute.manager [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Refreshing instance network info cache due to event network-changed-64ce8844-4f88-487b-8fba-e00413a3eb05. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.725780] env[61824]: DEBUG oslo_concurrency.lockutils [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] Acquiring lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.725978] env[61824]: DEBUG oslo_concurrency.lockutils [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] Acquired lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.726183] env[61824]: DEBUG nova.network.neutron [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Refreshing network info cache for port 64ce8844-4f88-487b-8fba-e00413a3eb05 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.076396] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275415, 'name': CreateVM_Task, 'duration_secs': 0.370487} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.076620] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 769.077573] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.077744] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.078134] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 769.078425] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e574d2df-c8bb-422f-ae7d-366f8e3fcf95 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.083520] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 769.083520] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]522149bf-3ffe-26d2-5776-055721f1f7f7" [ 769.083520] env[61824]: _type = "Task" [ 769.083520] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.092115] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522149bf-3ffe-26d2-5776-055721f1f7f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.133293] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275416, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481668} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.133592] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 769.133830] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.134119] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9cac9f88-9fde-46c8-9068-7838c46a6b99 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.141041] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 769.141041] env[61824]: value = "task-1275417" [ 769.141041] env[61824]: _type = "Task" [ 769.141041] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.147948] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275417, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.165864] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 6235a9b9-19c0-4917-938a-bca82a2ca05e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.411948] env[61824]: DEBUG nova.network.neutron [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Updated VIF entry in instance network info cache for port 64ce8844-4f88-487b-8fba-e00413a3eb05. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 769.412319] env[61824]: DEBUG nova.network.neutron [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Updating instance_info_cache with network_info: [{"id": "64ce8844-4f88-487b-8fba-e00413a3eb05", "address": "fa:16:3e:05:3e:0f", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64ce8844-4f", "ovs_interfaceid": "64ce8844-4f88-487b-8fba-e00413a3eb05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.593440] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522149bf-3ffe-26d2-5776-055721f1f7f7, 'name': SearchDatastore_Task, 'duration_secs': 0.009086} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.593713] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.593940] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 769.594187] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.594334] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.594508] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 769.594750] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-244ccb94-f76a-46ee-9f8e-27dc19b9ebd4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.602398] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 769.602565] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 769.603341] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b78141c1-eef2-414e-90ec-714ed0d2f9c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.607826] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 769.607826] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]524e17c0-b797-8a91-b548-99a8469c268b" [ 769.607826] env[61824]: _type = "Task" [ 769.607826] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.614607] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524e17c0-b797-8a91-b548-99a8469c268b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.649830] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275417, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076819} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.650119] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 769.650898] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e9e995-39ee-4be6-8deb-eb30ddd2a261 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.672530] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.673246] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 0201e783-ed75-4cee-8df4-efac97b4ca19 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.674405] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c313270-6ee8-45e3-a265-83137ec7b348 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.694714] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 769.694714] env[61824]: value = "task-1275418" [ 769.694714] env[61824]: _type = "Task" [ 769.694714] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.702362] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275418, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.915249] env[61824]: DEBUG oslo_concurrency.lockutils [req-c27b3b9b-7dd9-4914-83e4-473378b3b0ea req-be243f3b-d220-4502-9a70-5d752011ff8b service nova] Releasing lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.118702] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524e17c0-b797-8a91-b548-99a8469c268b, 'name': SearchDatastore_Task, 'duration_secs': 0.008566} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.119612] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6758d3d2-59fb-4032-9442-83d09bf45168 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.125760] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 770.125760] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]527f1e06-f286-b71d-1f5a-587efb0d9ac5" [ 770.125760] env[61824]: _type = "Task" [ 770.125760] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.133347] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527f1e06-f286-b71d-1f5a-587efb0d9ac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.190396] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance d9bb3172-c058-4e59-a719-d8aa340fba71 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.205627] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275418, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.636652] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527f1e06-f286-b71d-1f5a-587efb0d9ac5, 'name': SearchDatastore_Task, 'duration_secs': 0.010345} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.636985] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.637321] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 7f6a4446-86e1-44c7-ab42-297b033dbace/7f6a4446-86e1-44c7-ab42-297b033dbace.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 770.637600] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd3aaf8c-259a-484a-864b-b2f2dd4b3245 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.643520] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 770.643520] env[61824]: value = "task-1275419" [ 770.643520] env[61824]: _type = "Task" [ 770.643520] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.650816] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275419, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.693054] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7be79422-c624-4b4a-884f-29b13b9d81d2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.705485] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275418, 'name': ReconfigVM_Task, 'duration_secs': 0.510597} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.706418] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.707047] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c506282-6819-40d5-9105-11989b5ebb76 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.713263] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 770.713263] env[61824]: value = "task-1275420" [ 770.713263] env[61824]: _type = "Task" [ 770.713263] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.723588] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275420, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.153963] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275419, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45452} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.154259] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 7f6a4446-86e1-44c7-ab42-297b033dbace/7f6a4446-86e1-44c7-ab42-297b033dbace.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 771.154470] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 771.154711] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f74cb6e8-ec26-42bc-9177-ed2c74de375f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.161441] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 771.161441] env[61824]: value = "task-1275421" [ 771.161441] env[61824]: _type = "Task" [ 771.161441] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.168521] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275421, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.196454] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 43071be0-eaab-415c-8dcb-948bd9702885 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.222578] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275420, 'name': Rename_Task, 'duration_secs': 0.176507} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.222861] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 771.223124] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-65ed2907-22a8-4e82-ac00-e584928e12bf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.228356] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 771.228356] env[61824]: value = "task-1275422" [ 771.228356] env[61824]: _type = "Task" [ 771.228356] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.240832] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.671414] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275421, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061595} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.671774] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.672440] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e9fdc5-4b81-421a-bc48-f5afe5e23254 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.694165] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 7f6a4446-86e1-44c7-ab42-297b033dbace/7f6a4446-86e1-44c7-ab42-297b033dbace.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.694417] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57ef22c0-fb15-44cc-8a6e-889667e29092 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.709081] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance cbded637-56b9-4048-b66e-1edf6311f435 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.715755] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 771.715755] env[61824]: value = "task-1275423" [ 771.715755] env[61824]: _type = "Task" [ 771.715755] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.724268] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275423, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.737825] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275422, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.211765] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 2495fcb4-28cb-49c3-90fe-c84072466287 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.212167] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 772.212211] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 772.226218] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275423, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.237250] env[61824]: DEBUG oslo_vmware.api [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275422, 'name': PowerOnVM_Task, 'duration_secs': 0.536002} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.237586] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.237732] env[61824]: INFO nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Took 7.93 seconds to spawn the instance on the hypervisor. [ 772.237872] env[61824]: DEBUG nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.240860] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90dd2c67-133c-4a09-bcdd-71571b1250c2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.464596] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e50fa8-d0f1-4fad-b113-d60f6ca3606d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.473797] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a0d3b1-bd7c-4003-a62d-fa9d3ead0cc4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.503640] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd8d608-bbf6-4598-8c72-daf47dd9dbb9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.511256] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f090b5-e748-44bf-aa7f-8e297bdf0353 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.524635] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.727061] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275423, 'name': ReconfigVM_Task, 'duration_secs': 0.634667} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.727394] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 7f6a4446-86e1-44c7-ab42-297b033dbace/7f6a4446-86e1-44c7-ab42-297b033dbace.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.728029] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d6d2047-ffe4-4703-acd0-494cc5583081 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.733882] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 772.733882] env[61824]: value = "task-1275424" [ 772.733882] env[61824]: _type = "Task" [ 772.733882] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.741085] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275424, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.756563] env[61824]: INFO nova.compute.manager [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Took 29.61 seconds to build instance. [ 773.027768] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.244316] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275424, 'name': Rename_Task, 'duration_secs': 0.135468} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.244615] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.244824] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-247d9573-8c2b-42dc-a798-4c0d9a146431 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.250913] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 773.250913] env[61824]: value = "task-1275425" [ 773.250913] env[61824]: _type = "Task" [ 773.250913] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.259301] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc911b6a-b16c-419c-a073-cc64bd4e6218 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.138s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.259514] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275425, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.533516] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 773.533788] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.929s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.534096] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.038s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.536121] env[61824]: INFO nova.compute.claims [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.538758] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 773.538919] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Cleaning up deleted instances {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 773.760356] env[61824]: DEBUG oslo_vmware.api [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275425, 'name': PowerOnVM_Task, 'duration_secs': 0.474732} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.760749] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.760975] env[61824]: INFO nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Took 7.13 seconds to spawn the instance on the hypervisor. [ 773.761164] env[61824]: DEBUG nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 773.761935] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27484bbe-db5d-46c5-9cfc-8c24ecb3d057 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.764636] env[61824]: DEBUG nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 774.044625] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] There are 4 instances to clean {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 774.044908] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e758c89f-e5e2-42e0-926e-c629cb496a89] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 774.279920] env[61824]: INFO nova.compute.manager [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Took 29.28 seconds to build instance. [ 774.291625] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.551574] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: de5a46be-193a-40d8-aeba-d5d4de24ef95] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 774.769679] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07623324-abac-45a6-b1e8-6fcf32542d8f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.777685] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7efaadf-60c5-49a0-8249-07900063c9f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.782327] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3f5bf69-1b4c-485d-ade5-5ebb54e99b85 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.260s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.810257] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95c93c5-9ed4-43d9-b01c-1edbb38cfb9b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.817878] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4336955a-15bb-4059-9616-1749056d0cf2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.831357] env[61824]: DEBUG nova.compute.provider_tree [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.054362] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: ff1dac51-f008-4218-bd4c-f0104657a2ee] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 775.310587] env[61824]: DEBUG nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 775.336602] env[61824]: DEBUG nova.scheduler.client.report [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.557914] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 90a21ec5-7ccf-4d37-9ec9-19ba2bfff468] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 775.808828] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.809306] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.841408] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.841898] env[61824]: DEBUG nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 775.844307] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.570s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.925265] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.060941] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.061146] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Cleaning up deleted instances with incomplete migration {{(pid=61824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 776.349872] env[61824]: DEBUG nova.compute.utils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.354844] env[61824]: DEBUG nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 776.355047] env[61824]: DEBUG nova.network.neutron [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 776.372573] env[61824]: DEBUG nova.scheduler.client.report [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Refreshing inventories for resource provider 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 776.388125] env[61824]: DEBUG nova.scheduler.client.report [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Updating ProviderTree inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 776.388404] env[61824]: DEBUG nova.compute.provider_tree [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 776.400929] env[61824]: DEBUG nova.scheduler.client.report [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Refreshing aggregate associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, aggregates: None {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 776.414841] env[61824]: DEBUG nova.policy [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '92083101ef5947a4bf3b16d376f9fa20', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1dd4e264d4e34f708e96131a41747e97', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.420290] env[61824]: DEBUG nova.scheduler.client.report [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Refreshing trait associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 776.563172] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.664392] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1a8ad6-5db9-45c6-ae3b-5895fc992b7b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.673628] env[61824]: DEBUG nova.network.neutron [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Successfully created port: 1bee6aa9-7173-447c-838e-b5ca6c3c705e {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.676516] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4dae90-9181-408b-8556-38fbc577a3ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.705949] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41fb3b2-a43b-41f6-bfa7-f5baa0c980dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.713503] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfd1000-04e0-40c6-9e21-0c7b1ecf37b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.727416] env[61824]: DEBUG nova.compute.provider_tree [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.858478] env[61824]: DEBUG nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.230799] env[61824]: DEBUG nova.scheduler.client.report [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.736551] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.736914] env[61824]: ERROR nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Traceback (most recent call last): [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self.driver.spawn(context, instance, image_meta, [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] vm_ref = self.build_virtual_machine(instance, [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.736914] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] for vif in network_info: [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] return self._sync_wrapper(fn, *args, **kwargs) [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self.wait() [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self[:] = self._gt.wait() [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] return self._exit_event.wait() [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] current.throw(*self._exc) [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.737378] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] result = function(*args, **kwargs) [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] return func(*args, **kwargs) [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] raise e [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] nwinfo = self.network_api.allocate_for_instance( [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] created_port_ids = self._update_ports_for_instance( [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] with excutils.save_and_reraise_exception(): [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] self.force_reraise() [ 777.737809] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] raise self.value [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] updated_port = self._update_port( [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] _ensure_no_port_binding_failure(port) [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] raise exception.PortBindingFailed(port_id=port['id']) [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] nova.exception.PortBindingFailed: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. [ 777.738262] env[61824]: ERROR nova.compute.manager [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] [ 777.738262] env[61824]: DEBUG nova.compute.utils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.738877] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.361s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.740309] env[61824]: INFO nova.compute.claims [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.742816] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Build of instance 34f72773-5950-4710-b3e8-5e5c4877f227 was re-scheduled: Binding failed for port 7be17eae-3159-4e87-b87f-6dbab37400f2, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.743241] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.743462] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.743604] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.743760] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 777.867920] env[61824]: DEBUG nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 777.899571] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.899804] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.899956] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.900146] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.900291] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.900434] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.900633] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.900788] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.900951] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.901123] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.901396] env[61824]: DEBUG nova.virt.hardware [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.902263] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026e816a-6016-4e84-9314-49a6f911e405 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.912177] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c972de-f130-4504-be3e-41e7d5920dec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.188680] env[61824]: DEBUG nova.compute.manager [req-c878b53d-e998-4d0a-839d-fd26111c576f req-f5109ae0-85db-4635-8f98-d0cb3877f810 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Received event network-vif-plugged-1bee6aa9-7173-447c-838e-b5ca6c3c705e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.188909] env[61824]: DEBUG oslo_concurrency.lockutils [req-c878b53d-e998-4d0a-839d-fd26111c576f req-f5109ae0-85db-4635-8f98-d0cb3877f810 service nova] Acquiring lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.189122] env[61824]: DEBUG oslo_concurrency.lockutils [req-c878b53d-e998-4d0a-839d-fd26111c576f req-f5109ae0-85db-4635-8f98-d0cb3877f810 service nova] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.189284] env[61824]: DEBUG oslo_concurrency.lockutils [req-c878b53d-e998-4d0a-839d-fd26111c576f req-f5109ae0-85db-4635-8f98-d0cb3877f810 service nova] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.189447] env[61824]: DEBUG nova.compute.manager [req-c878b53d-e998-4d0a-839d-fd26111c576f req-f5109ae0-85db-4635-8f98-d0cb3877f810 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] No waiting events found dispatching network-vif-plugged-1bee6aa9-7173-447c-838e-b5ca6c3c705e {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 778.189605] env[61824]: WARNING nova.compute.manager [req-c878b53d-e998-4d0a-839d-fd26111c576f req-f5109ae0-85db-4635-8f98-d0cb3877f810 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Received unexpected event network-vif-plugged-1bee6aa9-7173-447c-838e-b5ca6c3c705e for instance with vm_state building and task_state spawning. [ 778.269655] env[61824]: DEBUG nova.network.neutron [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Successfully updated port: 1bee6aa9-7173-447c-838e-b5ca6c3c705e {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 778.271344] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.352290] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.776137] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "refresh_cache-23fbfb1e-a716-4a73-a15d-16ff25690e0e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.776719] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquired lock "refresh_cache-23fbfb1e-a716-4a73-a15d-16ff25690e0e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.776719] env[61824]: DEBUG nova.network.neutron [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.854985] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-34f72773-5950-4710-b3e8-5e5c4877f227" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.855209] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.855392] env[61824]: DEBUG nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.855555] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 778.871322] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.989527] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e79404c1-45f7-4e20-a031-173cb8ac0de2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.997718] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c03f0e-5dc0-4b78-9b7b-37ed12603b75 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.029066] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccba8e00-a755-43a1-85ae-b17e1b056945 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.039297] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec7d4b0-d06d-4633-9767-95fe96c55022 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.050955] env[61824]: DEBUG nova.compute.provider_tree [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.312304] env[61824]: DEBUG nova.network.neutron [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.373564] env[61824]: DEBUG nova.network.neutron [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.467921] env[61824]: DEBUG nova.network.neutron [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Updating instance_info_cache with network_info: [{"id": "1bee6aa9-7173-447c-838e-b5ca6c3c705e", "address": "fa:16:3e:b8:f6:99", "network": {"id": "173134bf-ee5a-4e1e-8dd8-99b222dc1fbd", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-116402407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dd4e264d4e34f708e96131a41747e97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bee6aa9-71", "ovs_interfaceid": "1bee6aa9-7173-447c-838e-b5ca6c3c705e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.554958] env[61824]: DEBUG nova.scheduler.client.report [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.876488] env[61824]: INFO nova.compute.manager [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 34f72773-5950-4710-b3e8-5e5c4877f227] Took 1.02 seconds to deallocate network for instance. [ 779.971015] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Releasing lock "refresh_cache-23fbfb1e-a716-4a73-a15d-16ff25690e0e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.971365] env[61824]: DEBUG nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Instance network_info: |[{"id": "1bee6aa9-7173-447c-838e-b5ca6c3c705e", "address": "fa:16:3e:b8:f6:99", "network": {"id": "173134bf-ee5a-4e1e-8dd8-99b222dc1fbd", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-116402407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dd4e264d4e34f708e96131a41747e97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bee6aa9-71", "ovs_interfaceid": "1bee6aa9-7173-447c-838e-b5ca6c3c705e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 779.971775] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:f6:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd4020f51-6e46-4b73-a79e-9fe3fd51b917', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1bee6aa9-7173-447c-838e-b5ca6c3c705e', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.979387] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Creating folder: Project (1dd4e264d4e34f708e96131a41747e97). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.979968] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb3baef9-3729-40fa-ad27-ec3686a5d689 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.990555] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Created folder: Project (1dd4e264d4e34f708e96131a41747e97) in parent group-v274074. [ 779.990742] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Creating folder: Instances. Parent ref: group-v274107. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 779.990953] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e531c11-adc5-4107-aaa5-5975227d9b75 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.999154] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Created folder: Instances in parent group-v274107. [ 779.999374] env[61824]: DEBUG oslo.service.loopingcall [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.999588] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 779.999773] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a17fd314-1091-449e-8dd7-9eee07204089 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.016926] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.016926] env[61824]: value = "task-1275428" [ 780.016926] env[61824]: _type = "Task" [ 780.016926] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.023868] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275428, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.059305] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.320s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.060059] env[61824]: DEBUG nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.062217] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.628s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.258725] env[61824]: DEBUG nova.compute.manager [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Received event network-changed-1bee6aa9-7173-447c-838e-b5ca6c3c705e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.258915] env[61824]: DEBUG nova.compute.manager [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Refreshing instance network info cache due to event network-changed-1bee6aa9-7173-447c-838e-b5ca6c3c705e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 780.259163] env[61824]: DEBUG oslo_concurrency.lockutils [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] Acquiring lock "refresh_cache-23fbfb1e-a716-4a73-a15d-16ff25690e0e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.259310] env[61824]: DEBUG oslo_concurrency.lockutils [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] Acquired lock "refresh_cache-23fbfb1e-a716-4a73-a15d-16ff25690e0e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.259477] env[61824]: DEBUG nova.network.neutron [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Refreshing network info cache for port 1bee6aa9-7173-447c-838e-b5ca6c3c705e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 780.528510] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275428, 'name': CreateVM_Task, 'duration_secs': 0.33233} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.528676] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 780.529342] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.529506] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.529824] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 780.530085] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ef2e42-4035-4bef-bf8a-e8b1dd9eb2f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.534284] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 780.534284] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]520b356e-8dff-d5e8-e99f-5e3b54ca534a" [ 780.534284] env[61824]: _type = "Task" [ 780.534284] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.541893] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520b356e-8dff-d5e8-e99f-5e3b54ca534a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.566238] env[61824]: DEBUG nova.compute.utils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.567526] env[61824]: DEBUG nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.567692] env[61824]: DEBUG nova.network.neutron [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.610008] env[61824]: DEBUG nova.policy [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96cc17dfb4fe4a3fbe2f1ecce380b08f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a2acf890b76493da7d39ea24a06f668', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.812949] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8441161-d9f7-44ff-9ab2-f003e640d7e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.821011] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a78f011-fbac-4a92-bd0a-701196bbedc7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.852779] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12e6c96-e76a-456d-91db-003ff9a8a2a2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.868768] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ca4a43-5d83-4f5e-8bfe-363f5bb1ad10 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.881906] env[61824]: DEBUG nova.compute.provider_tree [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.912433] env[61824]: INFO nova.scheduler.client.report [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleted allocations for instance 34f72773-5950-4710-b3e8-5e5c4877f227 [ 780.946337] env[61824]: DEBUG nova.network.neutron [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Successfully created port: cc85d94a-265d-4d8a-ade8-7ef62dfd117a {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.036241] env[61824]: DEBUG nova.network.neutron [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Updated VIF entry in instance network info cache for port 1bee6aa9-7173-447c-838e-b5ca6c3c705e. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.036595] env[61824]: DEBUG nova.network.neutron [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Updating instance_info_cache with network_info: [{"id": "1bee6aa9-7173-447c-838e-b5ca6c3c705e", "address": "fa:16:3e:b8:f6:99", "network": {"id": "173134bf-ee5a-4e1e-8dd8-99b222dc1fbd", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-116402407-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1dd4e264d4e34f708e96131a41747e97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d4020f51-6e46-4b73-a79e-9fe3fd51b917", "external-id": "nsx-vlan-transportzone-16", "segmentation_id": 16, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bee6aa9-71", "ovs_interfaceid": "1bee6aa9-7173-447c-838e-b5ca6c3c705e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.049016] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520b356e-8dff-d5e8-e99f-5e3b54ca534a, 'name': SearchDatastore_Task, 'duration_secs': 0.009213} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.049201] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.049391] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.049649] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.049788] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.049954] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.050433] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6eeeab9-e832-45de-b79a-833a41c10c0e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.058361] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.058546] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.059261] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6af6a61e-c392-43cb-b281-f1c107eebaa1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.064210] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 781.064210] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5246e861-fd7d-9d39-925c-69b24659fda2" [ 781.064210] env[61824]: _type = "Task" [ 781.064210] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.071453] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5246e861-fd7d-9d39-925c-69b24659fda2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.073023] env[61824]: DEBUG nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.392664] env[61824]: DEBUG nova.scheduler.client.report [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.421748] env[61824]: DEBUG oslo_concurrency.lockutils [None req-96e4d240-f2d7-4c53-8a44-2f5b81f822fb tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "34f72773-5950-4710-b3e8-5e5c4877f227" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.943s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.541734] env[61824]: DEBUG oslo_concurrency.lockutils [req-131f9889-e0f3-401f-a371-29abc94a1fb0 req-4726c55b-4675-4103-9f31-2e4535057738 service nova] Releasing lock "refresh_cache-23fbfb1e-a716-4a73-a15d-16ff25690e0e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.575867] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5246e861-fd7d-9d39-925c-69b24659fda2, 'name': SearchDatastore_Task, 'duration_secs': 0.007544} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.580589] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bb66f63-cbf5-430c-812b-c673733c9250 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.585876] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 781.585876] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]524e5ab1-c632-5235-6190-fc99e73c7b08" [ 781.585876] env[61824]: _type = "Task" [ 781.585876] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.593722] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524e5ab1-c632-5235-6190-fc99e73c7b08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.899466] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.837s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.900166] env[61824]: ERROR nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Traceback (most recent call last): [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self.driver.spawn(context, instance, image_meta, [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] vm_ref = self.build_virtual_machine(instance, [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.900166] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] for vif in network_info: [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] return self._sync_wrapper(fn, *args, **kwargs) [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self.wait() [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self[:] = self._gt.wait() [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] return self._exit_event.wait() [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] current.throw(*self._exc) [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.900719] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] result = function(*args, **kwargs) [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] return func(*args, **kwargs) [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] raise e [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] nwinfo = self.network_api.allocate_for_instance( [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] created_port_ids = self._update_ports_for_instance( [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] with excutils.save_and_reraise_exception(): [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] self.force_reraise() [ 781.901247] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] raise self.value [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] updated_port = self._update_port( [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] _ensure_no_port_binding_failure(port) [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] raise exception.PortBindingFailed(port_id=port['id']) [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] nova.exception.PortBindingFailed: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. [ 781.901684] env[61824]: ERROR nova.compute.manager [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] [ 781.901684] env[61824]: DEBUG nova.compute.utils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. {{(pid=61824) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.902057] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.808s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.903857] env[61824]: INFO nova.compute.claims [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.906422] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Build of instance 0f86f95d-2c41-47b2-93f8-c7e6983c913a was re-scheduled: Binding failed for port 7cabf642-da4b-4493-afc6-87b23fe52d99, please check neutron logs for more information. {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 781.906821] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Unplugging VIFs for instance {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 781.907050] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Acquiring lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.907211] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Acquired lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.907464] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.925781] env[61824]: DEBUG nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 782.081781] env[61824]: DEBUG nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.096321] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524e5ab1-c632-5235-6190-fc99e73c7b08, 'name': SearchDatastore_Task, 'duration_secs': 0.011489} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.096586] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.096837] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 23fbfb1e-a716-4a73-a15d-16ff25690e0e/23fbfb1e-a716-4a73-a15d-16ff25690e0e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 782.097101] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb400dbf-54f7-4e8a-ab4a-b23d82d933d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.103937] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 782.103937] env[61824]: value = "task-1275429" [ 782.103937] env[61824]: _type = "Task" [ 782.103937] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.112634] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275429, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.126415] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.126415] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.126415] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.126811] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.126811] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.126928] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.127125] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.127330] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.127666] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.127666] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.127833] env[61824]: DEBUG nova.virt.hardware [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.128637] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babec0e6-283b-44d5-a8b6-ddb90b5cc975 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.135686] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2c5a14-c6ea-4391-9aa5-7a59bebac8f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.427633] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.451250] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.513436] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.367140] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Releasing lock "refresh_cache-0f86f95d-2c41-47b2-93f8-c7e6983c913a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.367436] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61824) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 783.367564] env[61824]: DEBUG nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.367721] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.370382] env[61824]: DEBUG nova.compute.manager [req-f642ade6-1882-4e3d-9865-d70726aa277e req-12505038-03eb-4217-8d78-a9702db795aa service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Received event network-vif-plugged-cc85d94a-265d-4d8a-ade8-7ef62dfd117a {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.370649] env[61824]: DEBUG oslo_concurrency.lockutils [req-f642ade6-1882-4e3d-9865-d70726aa277e req-12505038-03eb-4217-8d78-a9702db795aa service nova] Acquiring lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.370847] env[61824]: DEBUG oslo_concurrency.lockutils [req-f642ade6-1882-4e3d-9865-d70726aa277e req-12505038-03eb-4217-8d78-a9702db795aa service nova] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.370966] env[61824]: DEBUG oslo_concurrency.lockutils [req-f642ade6-1882-4e3d-9865-d70726aa277e req-12505038-03eb-4217-8d78-a9702db795aa service nova] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.371127] env[61824]: DEBUG nova.compute.manager [req-f642ade6-1882-4e3d-9865-d70726aa277e req-12505038-03eb-4217-8d78-a9702db795aa service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] No waiting events found dispatching network-vif-plugged-cc85d94a-265d-4d8a-ade8-7ef62dfd117a {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 783.371284] env[61824]: WARNING nova.compute.manager [req-f642ade6-1882-4e3d-9865-d70726aa277e req-12505038-03eb-4217-8d78-a9702db795aa service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Received unexpected event network-vif-plugged-cc85d94a-265d-4d8a-ade8-7ef62dfd117a for instance with vm_state building and task_state spawning. [ 783.377867] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275429, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447864} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.378111] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 23fbfb1e-a716-4a73-a15d-16ff25690e0e/23fbfb1e-a716-4a73-a15d-16ff25690e0e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.380678] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.380678] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19ad35c3-cac2-4638-a974-26fcdf238726 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.386657] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 783.386657] env[61824]: value = "task-1275430" [ 783.386657] env[61824]: _type = "Task" [ 783.386657] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.389861] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.392124] env[61824]: DEBUG nova.network.neutron [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Successfully updated port: cc85d94a-265d-4d8a-ade8-7ef62dfd117a {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.396065] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275430, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.558491] env[61824]: DEBUG nova.compute.manager [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Received event network-changed-cc85d94a-265d-4d8a-ade8-7ef62dfd117a {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.558680] env[61824]: DEBUG nova.compute.manager [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Refreshing instance network info cache due to event network-changed-cc85d94a-265d-4d8a-ade8-7ef62dfd117a. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 783.558882] env[61824]: DEBUG oslo_concurrency.lockutils [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] Acquiring lock "refresh_cache-cc13d6a5-6b99-40bf-8978-09d284b4ed17" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.559030] env[61824]: DEBUG oslo_concurrency.lockutils [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] Acquired lock "refresh_cache-cc13d6a5-6b99-40bf-8978-09d284b4ed17" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.559191] env[61824]: DEBUG nova.network.neutron [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Refreshing network info cache for port cc85d94a-265d-4d8a-ade8-7ef62dfd117a {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 783.618853] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b38423-c95e-4ee5-a7a8-91ac956db47f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.626633] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3269b6-17b3-4a47-8f70-dc4b7a7a66d4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.656720] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e932352-fa3b-4df2-a653-0a8e9deb2cc6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.663845] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6883d4e8-de7e-4e66-8a93-bbc798faaa2f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.677821] env[61824]: DEBUG nova.compute.provider_tree [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.861998] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.862241] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.897514] env[61824]: DEBUG nova.network.neutron [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.899406] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "refresh_cache-cc13d6a5-6b99-40bf-8978-09d284b4ed17" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.899529] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275430, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069919} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.899985] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.901067] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075fdffb-0679-4ff2-ac79-582f8751ed67 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.934039] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 23fbfb1e-a716-4a73-a15d-16ff25690e0e/23fbfb1e-a716-4a73-a15d-16ff25690e0e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.934911] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff949c06-446e-407b-8666-c07985f79384 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.963188] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 783.963188] env[61824]: value = "task-1275431" [ 783.963188] env[61824]: _type = "Task" [ 783.963188] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.973563] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275431, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.100915] env[61824]: DEBUG nova.network.neutron [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 784.181011] env[61824]: DEBUG nova.scheduler.client.report [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 784.185304] env[61824]: DEBUG nova.network.neutron [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.400870] env[61824]: INFO nova.compute.manager [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] [instance: 0f86f95d-2c41-47b2-93f8-c7e6983c913a] Took 1.03 seconds to deallocate network for instance. [ 784.472445] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275431, 'name': ReconfigVM_Task, 'duration_secs': 0.257354} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.472710] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 23fbfb1e-a716-4a73-a15d-16ff25690e0e/23fbfb1e-a716-4a73-a15d-16ff25690e0e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.473317] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f0bc4d7a-4a17-4f6a-8239-776e4e23dfea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.479578] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 784.479578] env[61824]: value = "task-1275432" [ 784.479578] env[61824]: _type = "Task" [ 784.479578] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.487098] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275432, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.687389] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.785s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.687983] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.690695] env[61824]: DEBUG oslo_concurrency.lockutils [req-11104e8f-1fa3-425e-8757-ae59071f28ca req-c73b23ff-256c-4bbd-8a95-d0be8d4bac93 service nova] Releasing lock "refresh_cache-cc13d6a5-6b99-40bf-8978-09d284b4ed17" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.691515] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.774s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.692540] env[61824]: INFO nova.compute.claims [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.694701] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquired lock "refresh_cache-cc13d6a5-6b99-40bf-8978-09d284b4ed17" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.694858] env[61824]: DEBUG nova.network.neutron [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.989448] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275432, 'name': Rename_Task, 'duration_secs': 0.125158} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.991030] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 784.991030] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f328f8f3-8054-4a86-9cd0-046efb75e29b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.996499] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 784.996499] env[61824]: value = "task-1275433" [ 784.996499] env[61824]: _type = "Task" [ 784.996499] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.003677] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.192987] env[61824]: DEBUG nova.compute.utils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.194462] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 785.194634] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.227724] env[61824]: DEBUG nova.network.neutron [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.244881] env[61824]: DEBUG nova.policy [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a849332813d74fdb9d374eb0547efa0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1393b3c0605c4f139bf37f80d85bbe4d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.360983] env[61824]: DEBUG nova.network.neutron [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Updating instance_info_cache with network_info: [{"id": "cc85d94a-265d-4d8a-ade8-7ef62dfd117a", "address": "fa:16:3e:2a:fa:1b", "network": {"id": "5dc02cdc-8336-4318-8ad6-a3b8fdf571c8", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1521571127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a2acf890b76493da7d39ea24a06f668", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc85d94a-26", "ovs_interfaceid": "cc85d94a-265d-4d8a-ade8-7ef62dfd117a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.433427] env[61824]: INFO nova.scheduler.client.report [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Deleted allocations for instance 0f86f95d-2c41-47b2-93f8-c7e6983c913a [ 785.506346] env[61824]: DEBUG oslo_vmware.api [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275433, 'name': PowerOnVM_Task, 'duration_secs': 0.456127} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.506681] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 785.506897] env[61824]: INFO nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Took 7.64 seconds to spawn the instance on the hypervisor. [ 785.507098] env[61824]: DEBUG nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.508141] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb33c720-f120-4080-921e-ffe948e6e159 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.517599] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Successfully created port: 61a97b86-36a1-41d7-b3af-38ba07db954d {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.697531] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.863506] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Releasing lock "refresh_cache-cc13d6a5-6b99-40bf-8978-09d284b4ed17" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.863823] env[61824]: DEBUG nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Instance network_info: |[{"id": "cc85d94a-265d-4d8a-ade8-7ef62dfd117a", "address": "fa:16:3e:2a:fa:1b", "network": {"id": "5dc02cdc-8336-4318-8ad6-a3b8fdf571c8", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1521571127-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0a2acf890b76493da7d39ea24a06f668", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcc85d94a-26", "ovs_interfaceid": "cc85d94a-265d-4d8a-ade8-7ef62dfd117a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 785.866228] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:fa:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cc85d94a-265d-4d8a-ade8-7ef62dfd117a', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.873649] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Creating folder: Project (0a2acf890b76493da7d39ea24a06f668). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.874074] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48e51c9e-8405-4f45-bd35-f5bae3c2c07a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.886744] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Created folder: Project (0a2acf890b76493da7d39ea24a06f668) in parent group-v274074. [ 785.886744] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Creating folder: Instances. Parent ref: group-v274110. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.886956] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e312f63-47d4-4f90-9c09-3c99bd4eda13 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.895353] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Created folder: Instances in parent group-v274110. [ 785.895602] env[61824]: DEBUG oslo.service.loopingcall [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.897823] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.898539] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-639e50c3-142a-47a9-9bc0-9b8d1b5d0497 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.919470] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.919470] env[61824]: value = "task-1275436" [ 785.919470] env[61824]: _type = "Task" [ 785.919470] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.927798] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275436, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.942698] env[61824]: DEBUG oslo_concurrency.lockutils [None req-54ecd28f-e203-4d11-8175-414c11bcde82 tempest-ServerActionsTestJSON-246940483 tempest-ServerActionsTestJSON-246940483-project-member] Lock "0f86f95d-2c41-47b2-93f8-c7e6983c913a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.094s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.952785] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0828fa-bb73-4d9c-b848-830ae2b027ba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.961428] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4fbe02-a7f6-45fc-978d-a7cf54a4a52a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.993405] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fd0751-cf46-4160-832d-2d953e13727f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.001109] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab2ad2af-1939-4b8e-9a38-a45b20fa82a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.019864] env[61824]: DEBUG nova.compute.provider_tree [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.029365] env[61824]: INFO nova.compute.manager [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Took 33.55 seconds to build instance. [ 786.429974] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275436, 'name': CreateVM_Task, 'duration_secs': 0.286666} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.430194] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.430899] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.431072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.431405] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 786.431649] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31a1080d-4779-4ebd-9ba4-19f19ad375db {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.436333] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 786.436333] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52670c3f-568c-3478-6410-ec1ede0915f9" [ 786.436333] env[61824]: _type = "Task" [ 786.436333] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.443867] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52670c3f-568c-3478-6410-ec1ede0915f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.445332] env[61824]: DEBUG nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.525826] env[61824]: DEBUG nova.scheduler.client.report [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.532872] env[61824]: DEBUG oslo_concurrency.lockutils [None req-694fe143-292c-4b2d-b96e-f9c425f14502 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.860s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.709672] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.733872] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.734142] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.734297] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.734477] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.734701] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.734761] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.734957] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.735130] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.735313] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.735486] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.735632] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.736483] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3917c8-1852-4df3-9161-047c0341b81a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.744563] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d677dfde-9a73-4bf4-a918-efd41d95991f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.947161] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52670c3f-568c-3478-6410-ec1ede0915f9, 'name': SearchDatastore_Task, 'duration_secs': 0.008401} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.947486] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.947718] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.947945] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.948106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.948296] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.950835] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a92f718-e5c2-4ee1-b7ba-1489f4fc813d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.961324] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.961324] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 786.961324] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d752fcbb-4844-49c6-be6f-d43e87395f5b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.970634] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 786.970634] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52b25580-cfbb-ddca-5a3a-ba149846aebd" [ 786.970634] env[61824]: _type = "Task" [ 786.970634] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.975041] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.978450] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b25580-cfbb-ddca-5a3a-ba149846aebd, 'name': SearchDatastore_Task, 'duration_secs': 0.007605} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.979169] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-667f55ed-25c5-4837-9d02-e711768fe86d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.984090] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 786.984090] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]528db3b5-f6ba-5f89-721f-5e487b617f2d" [ 786.984090] env[61824]: _type = "Task" [ 786.984090] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.994092] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528db3b5-f6ba-5f89-721f-5e487b617f2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.031243] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.031342] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 787.033875] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.300s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.035268] env[61824]: INFO nova.compute.claims [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.037831] env[61824]: DEBUG nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.495175] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528db3b5-f6ba-5f89-721f-5e487b617f2d, 'name': SearchDatastore_Task, 'duration_secs': 0.037857} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.495593] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.495709] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] cc13d6a5-6b99-40bf-8978-09d284b4ed17/cc13d6a5-6b99-40bf-8978-09d284b4ed17.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 787.495949] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa32794e-5660-44cb-99c3-0b415306f5b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.502408] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 787.502408] env[61824]: value = "task-1275437" [ 787.502408] env[61824]: _type = "Task" [ 787.502408] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.510383] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275437, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.539439] env[61824]: DEBUG nova.compute.utils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.541168] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.541404] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 787.562108] env[61824]: DEBUG nova.compute.manager [req-921c334e-f8f3-4fcb-8e12-fa46ae3218b8 req-8e016d91-9472-43a1-9751-b68d1e58258d service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Received event network-vif-plugged-61a97b86-36a1-41d7-b3af-38ba07db954d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.562421] env[61824]: DEBUG oslo_concurrency.lockutils [req-921c334e-f8f3-4fcb-8e12-fa46ae3218b8 req-8e016d91-9472-43a1-9751-b68d1e58258d service nova] Acquiring lock "f24881d3-a647-42f9-bf82-0065f3425a5f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.562680] env[61824]: DEBUG oslo_concurrency.lockutils [req-921c334e-f8f3-4fcb-8e12-fa46ae3218b8 req-8e016d91-9472-43a1-9751-b68d1e58258d service nova] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.562858] env[61824]: DEBUG oslo_concurrency.lockutils [req-921c334e-f8f3-4fcb-8e12-fa46ae3218b8 req-8e016d91-9472-43a1-9751-b68d1e58258d service nova] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.563037] env[61824]: DEBUG nova.compute.manager [req-921c334e-f8f3-4fcb-8e12-fa46ae3218b8 req-8e016d91-9472-43a1-9751-b68d1e58258d service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] No waiting events found dispatching network-vif-plugged-61a97b86-36a1-41d7-b3af-38ba07db954d {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 787.563202] env[61824]: WARNING nova.compute.manager [req-921c334e-f8f3-4fcb-8e12-fa46ae3218b8 req-8e016d91-9472-43a1-9751-b68d1e58258d service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Received unexpected event network-vif-plugged-61a97b86-36a1-41d7-b3af-38ba07db954d for instance with vm_state building and task_state spawning. [ 787.572571] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.598034] env[61824]: DEBUG nova.policy [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a849332813d74fdb9d374eb0547efa0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1393b3c0605c4f139bf37f80d85bbe4d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.885281] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Successfully created port: 47dcce78-bb07-4111-9322-e92f8f6763b1 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 788.013633] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275437, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447079} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.014071] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] cc13d6a5-6b99-40bf-8978-09d284b4ed17/cc13d6a5-6b99-40bf-8978-09d284b4ed17.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 788.014434] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.015411] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Successfully updated port: 61a97b86-36a1-41d7-b3af-38ba07db954d {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 788.017059] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0035451d-0a0b-4ca8-8d49-0e1a68070271 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.027651] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 788.027651] env[61824]: value = "task-1275438" [ 788.027651] env[61824]: _type = "Task" [ 788.027651] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.041487] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275438, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.044049] env[61824]: DEBUG nova.compute.manager [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Received event network-changed-61a97b86-36a1-41d7-b3af-38ba07db954d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.044290] env[61824]: DEBUG nova.compute.manager [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Refreshing instance network info cache due to event network-changed-61a97b86-36a1-41d7-b3af-38ba07db954d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.044562] env[61824]: DEBUG oslo_concurrency.lockutils [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] Acquiring lock "refresh_cache-f24881d3-a647-42f9-bf82-0065f3425a5f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.044767] env[61824]: DEBUG oslo_concurrency.lockutils [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] Acquired lock "refresh_cache-f24881d3-a647-42f9-bf82-0065f3425a5f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.044966] env[61824]: DEBUG nova.network.neutron [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Refreshing network info cache for port 61a97b86-36a1-41d7-b3af-38ba07db954d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 788.047934] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.290668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.290668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.290668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.290668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.290989] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.292709] env[61824]: INFO nova.compute.manager [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Terminating instance [ 788.303032] env[61824]: DEBUG nova.compute.manager [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 788.303032] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.303032] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e184d643-21ec-4bf8-a242-b4ad7a8704cd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.309416] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 788.309674] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d519b3b-00f3-4a00-ab23-8d80e39a1cad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.317718] env[61824]: DEBUG oslo_vmware.api [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 788.317718] env[61824]: value = "task-1275439" [ 788.317718] env[61824]: _type = "Task" [ 788.317718] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.326349] env[61824]: DEBUG oslo_vmware.api [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.344886] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84d83db-afcd-44db-b053-52f7919e78c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.352237] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7ca39f-9142-4613-ab2c-f77e436c6305 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.385086] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be70163c-f4fa-4019-8400-458e9684f4e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.392470] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70be88c9-8c86-4efa-9a11-34ae385fab02 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.406973] env[61824]: DEBUG nova.compute.provider_tree [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.525170] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "refresh_cache-f24881d3-a647-42f9-bf82-0065f3425a5f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.538930] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275438, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069679} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.539349] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 788.540140] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934aa49f-a4eb-49bf-878a-cdc840cc8e7c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.570204] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] cc13d6a5-6b99-40bf-8978-09d284b4ed17/cc13d6a5-6b99-40bf-8978-09d284b4ed17.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.570691] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78791d12-1b99-45ac-b901-57f40a9dae10 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.591190] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 788.591190] env[61824]: value = "task-1275440" [ 788.591190] env[61824]: _type = "Task" [ 788.591190] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.599781] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275440, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.624785] env[61824]: DEBUG nova.network.neutron [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.784512] env[61824]: DEBUG nova.network.neutron [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.830792] env[61824]: DEBUG oslo_vmware.api [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275439, 'name': PowerOffVM_Task, 'duration_secs': 0.19937} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.830792] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 788.830792] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 788.830792] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99ef304f-ea74-4199-b2a7-525cb168a906 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.891242] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 788.891608] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 788.891888] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Deleting the datastore file [datastore2] 23fbfb1e-a716-4a73-a15d-16ff25690e0e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.892208] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-074d59c6-f3b1-40f8-96b1-982f94c7f907 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.899927] env[61824]: DEBUG oslo_vmware.api [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for the task: (returnval){ [ 788.899927] env[61824]: value = "task-1275442" [ 788.899927] env[61824]: _type = "Task" [ 788.899927] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.907992] env[61824]: DEBUG oslo_vmware.api [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275442, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.910036] env[61824]: DEBUG nova.scheduler.client.report [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.072015] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 789.099543] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 789.099802] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 789.099960] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 789.100156] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 789.100301] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 789.100446] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 789.100646] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 789.100800] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 789.100963] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 789.101137] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 789.101309] env[61824]: DEBUG nova.virt.hardware [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.102466] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55dce9a-def6-4896-b858-7db30784bd41 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.108434] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275440, 'name': ReconfigVM_Task, 'duration_secs': 0.25667} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.109247] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Reconfigured VM instance instance-00000038 to attach disk [datastore2] cc13d6a5-6b99-40bf-8978-09d284b4ed17/cc13d6a5-6b99-40bf-8978-09d284b4ed17.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.110240] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f3590823-001d-4a5f-87a8-cfa6643b2974 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.115786] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70925ee5-eaf1-4761-b071-8857d6070295 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.120687] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 789.120687] env[61824]: value = "task-1275443" [ 789.120687] env[61824]: _type = "Task" [ 789.120687] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.136395] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275443, 'name': Rename_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.289769] env[61824]: DEBUG oslo_concurrency.lockutils [req-bb2b674f-e9fe-4564-a190-147008e46649 req-955f5b92-f841-4a9d-be8f-45c9ab853019 service nova] Releasing lock "refresh_cache-f24881d3-a647-42f9-bf82-0065f3425a5f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.290233] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "refresh_cache-f24881d3-a647-42f9-bf82-0065f3425a5f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.291023] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 789.327973] env[61824]: DEBUG nova.compute.manager [req-49540299-3a96-44f7-83eb-936c13d55d0d req-ec060eb7-3594-48e8-b6e5-0f994b2e9a5d service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Received event network-vif-plugged-47dcce78-bb07-4111-9322-e92f8f6763b1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.328220] env[61824]: DEBUG oslo_concurrency.lockutils [req-49540299-3a96-44f7-83eb-936c13d55d0d req-ec060eb7-3594-48e8-b6e5-0f994b2e9a5d service nova] Acquiring lock "e3e19a04-a29b-4473-833c-d8486703e9ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.328544] env[61824]: DEBUG oslo_concurrency.lockutils [req-49540299-3a96-44f7-83eb-936c13d55d0d req-ec060eb7-3594-48e8-b6e5-0f994b2e9a5d service nova] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.328611] env[61824]: DEBUG oslo_concurrency.lockutils [req-49540299-3a96-44f7-83eb-936c13d55d0d req-ec060eb7-3594-48e8-b6e5-0f994b2e9a5d service nova] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.328789] env[61824]: DEBUG nova.compute.manager [req-49540299-3a96-44f7-83eb-936c13d55d0d req-ec060eb7-3594-48e8-b6e5-0f994b2e9a5d service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] No waiting events found dispatching network-vif-plugged-47dcce78-bb07-4111-9322-e92f8f6763b1 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 789.328932] env[61824]: WARNING nova.compute.manager [req-49540299-3a96-44f7-83eb-936c13d55d0d req-ec060eb7-3594-48e8-b6e5-0f994b2e9a5d service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Received unexpected event network-vif-plugged-47dcce78-bb07-4111-9322-e92f8f6763b1 for instance with vm_state building and task_state spawning. [ 789.410891] env[61824]: DEBUG oslo_vmware.api [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Task: {'id': task-1275442, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137359} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.410891] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.411091] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 789.411445] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.411445] env[61824]: INFO nova.compute.manager [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 789.411659] env[61824]: DEBUG oslo.service.loopingcall [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.411842] env[61824]: DEBUG nova.compute.manager [-] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.411933] env[61824]: DEBUG nova.network.neutron [-] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 789.417261] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Successfully updated port: 47dcce78-bb07-4111-9322-e92f8f6763b1 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 789.419229] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.419686] env[61824]: DEBUG nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.422161] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.530s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.422344] env[61824]: DEBUG nova.objects.instance [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 789.635938] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275443, 'name': Rename_Task, 'duration_secs': 0.14428} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.636321] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 789.636585] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf453783-d594-4999-a3a3-b3aeb76308fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.647859] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 789.647859] env[61824]: value = "task-1275444" [ 789.647859] env[61824]: _type = "Task" [ 789.647859] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.658930] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.829382] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 789.923178] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "refresh_cache-e3e19a04-a29b-4473-833c-d8486703e9ad" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.923340] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "refresh_cache-e3e19a04-a29b-4473-833c-d8486703e9ad" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.923499] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 789.926430] env[61824]: DEBUG nova.compute.utils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.934886] env[61824]: DEBUG nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.934986] env[61824]: DEBUG nova.network.neutron [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 790.025745] env[61824]: DEBUG nova.policy [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6a2f1ffe8934519acdcaca77e2a55d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e51dd2f578344aa8afa5d6beccb54f1e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.159246] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275444, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.170282] env[61824]: DEBUG nova.compute.manager [req-403d4d1b-5a17-4f74-b88f-580e6652fa0d req-05db5906-f2b5-40f4-9cd2-f2e2a270e11c service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Received event network-vif-deleted-1bee6aa9-7173-447c-838e-b5ca6c3c705e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 790.170516] env[61824]: INFO nova.compute.manager [req-403d4d1b-5a17-4f74-b88f-580e6652fa0d req-05db5906-f2b5-40f4-9cd2-f2e2a270e11c service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Neutron deleted interface 1bee6aa9-7173-447c-838e-b5ca6c3c705e; detaching it from the instance and deleting it from the info cache [ 790.170646] env[61824]: DEBUG nova.network.neutron [req-403d4d1b-5a17-4f74-b88f-580e6652fa0d req-05db5906-f2b5-40f4-9cd2-f2e2a270e11c service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.235975] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Updating instance_info_cache with network_info: [{"id": "61a97b86-36a1-41d7-b3af-38ba07db954d", "address": "fa:16:3e:49:be:38", "network": {"id": "c9b73dc1-0156-4446-a642-4fc913e57ed5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-199997673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1393b3c0605c4f139bf37f80d85bbe4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61a97b86-36", "ovs_interfaceid": "61a97b86-36a1-41d7-b3af-38ba07db954d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.262859] env[61824]: DEBUG nova.network.neutron [-] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.436024] env[61824]: DEBUG nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 790.438857] env[61824]: DEBUG oslo_concurrency.lockutils [None req-da5f7dff-e079-4be1-ae40-f4c63801dab6 tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.440085] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.802s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.441329] env[61824]: INFO nova.compute.claims [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.462842] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 790.502988] env[61824]: DEBUG nova.network.neutron [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Successfully created port: e95afb5b-11f8-4ed4-abb0-a12b48c9c14d {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.642223] env[61824]: DEBUG nova.network.neutron [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Updating instance_info_cache with network_info: [{"id": "47dcce78-bb07-4111-9322-e92f8f6763b1", "address": "fa:16:3e:09:c2:ee", "network": {"id": "c9b73dc1-0156-4446-a642-4fc913e57ed5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-199997673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1393b3c0605c4f139bf37f80d85bbe4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47dcce78-bb", "ovs_interfaceid": "47dcce78-bb07-4111-9322-e92f8f6763b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.659621] env[61824]: DEBUG oslo_vmware.api [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275444, 'name': PowerOnVM_Task, 'duration_secs': 0.851049} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.659906] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.660131] env[61824]: INFO nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Took 8.58 seconds to spawn the instance on the hypervisor. [ 790.660316] env[61824]: DEBUG nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.661720] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f136d28c-4945-44fc-a0c0-030db74761c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.672930] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a5a4167-0d4a-4eeb-9bba-90803e1febd0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.681495] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f9533f-b0a2-4fcb-8f62-914de07679b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.707968] env[61824]: DEBUG nova.compute.manager [req-403d4d1b-5a17-4f74-b88f-580e6652fa0d req-05db5906-f2b5-40f4-9cd2-f2e2a270e11c service nova] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Detach interface failed, port_id=1bee6aa9-7173-447c-838e-b5ca6c3c705e, reason: Instance 23fbfb1e-a716-4a73-a15d-16ff25690e0e could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 790.738658] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "refresh_cache-f24881d3-a647-42f9-bf82-0065f3425a5f" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.739097] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Instance network_info: |[{"id": "61a97b86-36a1-41d7-b3af-38ba07db954d", "address": "fa:16:3e:49:be:38", "network": {"id": "c9b73dc1-0156-4446-a642-4fc913e57ed5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-199997673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1393b3c0605c4f139bf37f80d85bbe4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61a97b86-36", "ovs_interfaceid": "61a97b86-36a1-41d7-b3af-38ba07db954d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 790.739683] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:be:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5c8dbe25-bca7-4d91-b577-193b8b2aad8d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61a97b86-36a1-41d7-b3af-38ba07db954d', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.747501] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Creating folder: Project (1393b3c0605c4f139bf37f80d85bbe4d). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.750490] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cdca9b06-24b4-4015-8e66-a1de81f54a08 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.758084] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Created folder: Project (1393b3c0605c4f139bf37f80d85bbe4d) in parent group-v274074. [ 790.758519] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Creating folder: Instances. Parent ref: group-v274113. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 790.758519] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-022d2c77-8a09-4916-a974-d13ff0c9220d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.765306] env[61824]: INFO nova.compute.manager [-] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Took 1.35 seconds to deallocate network for instance. [ 790.768569] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Created folder: Instances in parent group-v274113. [ 790.768733] env[61824]: DEBUG oslo.service.loopingcall [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.771620] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 790.772023] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-199e21b4-eb8d-490a-9416-f699c9b9447b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.793268] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.793268] env[61824]: value = "task-1275447" [ 790.793268] env[61824]: _type = "Task" [ 790.793268] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.804200] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275447, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.145521] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "refresh_cache-e3e19a04-a29b-4473-833c-d8486703e9ad" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.145827] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Instance network_info: |[{"id": "47dcce78-bb07-4111-9322-e92f8f6763b1", "address": "fa:16:3e:09:c2:ee", "network": {"id": "c9b73dc1-0156-4446-a642-4fc913e57ed5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-199997673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1393b3c0605c4f139bf37f80d85bbe4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47dcce78-bb", "ovs_interfaceid": "47dcce78-bb07-4111-9322-e92f8f6763b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 791.146248] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:c2:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5c8dbe25-bca7-4d91-b577-193b8b2aad8d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47dcce78-bb07-4111-9322-e92f8f6763b1', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 791.155092] env[61824]: DEBUG oslo.service.loopingcall [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.155461] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 791.155592] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa84f60d-847a-4c61-9843-bd3733af5a5d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.182206] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 791.182206] env[61824]: value = "task-1275448" [ 791.182206] env[61824]: _type = "Task" [ 791.182206] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.190516] env[61824]: INFO nova.compute.manager [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Took 35.82 seconds to build instance. [ 791.199839] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275448, 'name': CreateVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.273128] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.303937] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275447, 'name': CreateVM_Task, 'duration_secs': 0.335176} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.304247] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 791.304837] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.304999] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.305365] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.305631] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7994e24e-b191-4c44-b663-2061f4872292 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.310255] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 791.310255] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52bc9383-0364-c8c0-c653-6705422d91d9" [ 791.310255] env[61824]: _type = "Task" [ 791.310255] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.318728] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bc9383-0364-c8c0-c653-6705422d91d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.377139] env[61824]: DEBUG nova.compute.manager [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Received event network-changed-47dcce78-bb07-4111-9322-e92f8f6763b1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.377139] env[61824]: DEBUG nova.compute.manager [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Refreshing instance network info cache due to event network-changed-47dcce78-bb07-4111-9322-e92f8f6763b1. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 791.377139] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] Acquiring lock "refresh_cache-e3e19a04-a29b-4473-833c-d8486703e9ad" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.377683] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] Acquired lock "refresh_cache-e3e19a04-a29b-4473-833c-d8486703e9ad" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.378051] env[61824]: DEBUG nova.network.neutron [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Refreshing network info cache for port 47dcce78-bb07-4111-9322-e92f8f6763b1 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 791.451831] env[61824]: DEBUG nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 791.483325] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.483325] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.483325] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.485344] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.485344] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.485344] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.485344] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.485344] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.485518] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.485518] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.485518] env[61824]: DEBUG nova.virt.hardware [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.486142] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7476bb5d-ba57-4afb-af9a-21e53cf1bb70 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.498536] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-915bd13b-ef89-4db4-a8ca-60cad96a0529 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.693964] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6954062e-5eea-42ca-8cda-f5fffc33f8cf tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.350s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.694271] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275448, 'name': CreateVM_Task, 'duration_secs': 0.337105} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.697481] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 791.698273] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.746314] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-913c21ce-7102-4bd4-b2b8-acb79abd6493 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.755596] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62892679-9ecd-42db-9485-f417c9b907e3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.791947] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830e608a-a21f-4e3b-95e5-85076450463e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.799646] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd0c981-db80-45f8-a897-736b4df6752b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.814741] env[61824]: DEBUG nova.compute.provider_tree [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.825225] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bc9383-0364-c8c0-c653-6705422d91d9, 'name': SearchDatastore_Task, 'duration_secs': 0.00967} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.825534] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.825760] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.825987] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.826145] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.826561] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.827519] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.827844] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.828034] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93265961-0118-4436-bd5a-ec4da608d262 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.829705] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7358bbdb-7e24-4207-8633-9c18cb97bac1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.835339] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 791.835339] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52bab9ed-5da5-d40d-cb0d-5608f9ee73c6" [ 791.835339] env[61824]: _type = "Task" [ 791.835339] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.838898] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.839045] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 791.840274] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f1e1b93-49ce-483c-a94f-d031e616d19e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.845307] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bab9ed-5da5-d40d-cb0d-5608f9ee73c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.847950] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 791.847950] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5256eeab-5b64-654d-b51e-22c09efd9b1a" [ 791.847950] env[61824]: _type = "Task" [ 791.847950] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.855229] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5256eeab-5b64-654d-b51e-22c09efd9b1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.199886] env[61824]: DEBUG nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.274926] env[61824]: DEBUG nova.network.neutron [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Updated VIF entry in instance network info cache for port 47dcce78-bb07-4111-9322-e92f8f6763b1. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 792.275144] env[61824]: DEBUG nova.network.neutron [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Updating instance_info_cache with network_info: [{"id": "47dcce78-bb07-4111-9322-e92f8f6763b1", "address": "fa:16:3e:09:c2:ee", "network": {"id": "c9b73dc1-0156-4446-a642-4fc913e57ed5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-199997673-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1393b3c0605c4f139bf37f80d85bbe4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5c8dbe25-bca7-4d91-b577-193b8b2aad8d", "external-id": "nsx-vlan-transportzone-643", "segmentation_id": 643, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47dcce78-bb", "ovs_interfaceid": "47dcce78-bb07-4111-9322-e92f8f6763b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.320445] env[61824]: DEBUG nova.scheduler.client.report [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.352781] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bab9ed-5da5-d40d-cb0d-5608f9ee73c6, 'name': SearchDatastore_Task, 'duration_secs': 0.01051} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.352781] env[61824]: DEBUG nova.network.neutron [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Successfully updated port: e95afb5b-11f8-4ed4-abb0-a12b48c9c14d {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 792.357083] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.358298] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 792.358298] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.368269] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5256eeab-5b64-654d-b51e-22c09efd9b1a, 'name': SearchDatastore_Task, 'duration_secs': 0.008427} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.369082] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c66264e-1cd9-4bd1-b653-2095a1771fcd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.376957] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 792.376957] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a83a01-1be0-61a9-9289-438219de6e9a" [ 792.376957] env[61824]: _type = "Task" [ 792.376957] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.386192] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a83a01-1be0-61a9-9289-438219de6e9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.728785] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.780079] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb7dafd9-09ea-4635-8c2a-b7e82fadf643 req-1045c165-2967-43f4-90aa-8b7227fb83b9 service nova] Releasing lock "refresh_cache-e3e19a04-a29b-4473-833c-d8486703e9ad" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.826688] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.830326] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.585s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.830326] env[61824]: DEBUG nova.objects.instance [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lazy-loading 'resources' on Instance uuid 9fd8e236-3556-4b95-952f-3c324b896e29 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.859649] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.859928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.860592] env[61824]: DEBUG nova.network.neutron [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 792.888996] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a83a01-1be0-61a9-9289-438219de6e9a, 'name': SearchDatastore_Task, 'duration_secs': 0.011119} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.889266] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.889517] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] f24881d3-a647-42f9-bf82-0065f3425a5f/f24881d3-a647-42f9-bf82-0065f3425a5f.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 792.889791] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.889970] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 792.890191] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dfc9e18f-186a-4623-aa88-1ac74e3a05cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.892127] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d751fcc-65fe-4d6b-89d3-5fdcb2e27893 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.899981] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 792.899981] env[61824]: value = "task-1275449" [ 792.899981] env[61824]: _type = "Task" [ 792.899981] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.901617] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 792.901790] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 792.905603] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70b48a60-375e-4610-9bde-56a1957f8f7b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.914825] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275449, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.915159] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 792.915159] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]523b98c6-b296-ae8c-bb05-86fb8a72b9e2" [ 792.915159] env[61824]: _type = "Task" [ 792.915159] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.923386] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523b98c6-b296-ae8c-bb05-86fb8a72b9e2, 'name': SearchDatastore_Task, 'duration_secs': 0.008007} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.924518] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55c5acad-ce85-4830-864a-9daae656b9f4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.929561] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 792.929561] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1fe58-0f8b-9e52-9ab5-4fd3117b40e5" [ 792.929561] env[61824]: _type = "Task" [ 792.929561] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.937419] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1fe58-0f8b-9e52-9ab5-4fd3117b40e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.042437] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.042806] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.043060] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.043259] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.043471] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.045904] env[61824]: INFO nova.compute.manager [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Terminating instance [ 793.048353] env[61824]: DEBUG nova.compute.manager [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.048578] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.049487] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d949faf-9db9-41a4-9127-929b333cc0db {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.059815] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 793.063138] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8a7f95a4-fb62-4a82-996b-87b35551417e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.072155] env[61824]: DEBUG oslo_vmware.api [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 793.072155] env[61824]: value = "task-1275450" [ 793.072155] env[61824]: _type = "Task" [ 793.072155] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.082356] env[61824]: DEBUG oslo_vmware.api [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275450, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.333304] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "68d46a6e-7903-4b5b-a5c2-928a97cbe23e" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.333572] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "68d46a6e-7903-4b5b-a5c2-928a97cbe23e" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.412201] env[61824]: DEBUG nova.network.neutron [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.414988] env[61824]: DEBUG nova.compute.manager [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Received event network-vif-plugged-e95afb5b-11f8-4ed4-abb0-a12b48c9c14d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.415269] env[61824]: DEBUG oslo_concurrency.lockutils [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] Acquiring lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.415469] env[61824]: DEBUG oslo_concurrency.lockutils [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.415638] env[61824]: DEBUG oslo_concurrency.lockutils [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.415801] env[61824]: DEBUG nova.compute.manager [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] No waiting events found dispatching network-vif-plugged-e95afb5b-11f8-4ed4-abb0-a12b48c9c14d {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 793.415960] env[61824]: WARNING nova.compute.manager [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Received unexpected event network-vif-plugged-e95afb5b-11f8-4ed4-abb0-a12b48c9c14d for instance with vm_state building and task_state spawning. [ 793.416159] env[61824]: DEBUG nova.compute.manager [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Received event network-changed-e95afb5b-11f8-4ed4-abb0-a12b48c9c14d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.416326] env[61824]: DEBUG nova.compute.manager [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Refreshing instance network info cache due to event network-changed-e95afb5b-11f8-4ed4-abb0-a12b48c9c14d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.416490] env[61824]: DEBUG oslo_concurrency.lockutils [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] Acquiring lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.425484] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275449, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441119} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.425793] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] f24881d3-a647-42f9-bf82-0065f3425a5f/f24881d3-a647-42f9-bf82-0065f3425a5f.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 793.425916] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.426164] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8501a0b-6657-4c25-a19d-faf241b9e8ed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.442343] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 793.442343] env[61824]: value = "task-1275451" [ 793.442343] env[61824]: _type = "Task" [ 793.442343] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.452101] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1fe58-0f8b-9e52-9ab5-4fd3117b40e5, 'name': SearchDatastore_Task, 'duration_secs': 0.007699} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.454596] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.454856] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] e3e19a04-a29b-4473-833c-d8486703e9ad/e3e19a04-a29b-4473-833c-d8486703e9ad.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 793.455121] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2339ffa-6010-47dc-b659-32b56b8704e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.459951] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275451, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.465284] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 793.465284] env[61824]: value = "task-1275452" [ 793.465284] env[61824]: _type = "Task" [ 793.465284] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.473550] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275452, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.584418] env[61824]: DEBUG oslo_vmware.api [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275450, 'name': PowerOffVM_Task, 'duration_secs': 0.304735} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.584716] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 793.584905] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 793.589879] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3cf95b85-0a3f-4010-9a00-409895fac1f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.659053] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 793.659481] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 793.659814] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Deleting the datastore file [datastore2] cc13d6a5-6b99-40bf-8978-09d284b4ed17 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.661754] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ede3ab68-9c75-419f-98d1-79017d5488ad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.664820] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505f5f17-b35b-4dd7-a5bf-c156f33c527e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.676118] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b77e5f-b315-4582-bc17-09df2e4377d9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.680482] env[61824]: DEBUG oslo_vmware.api [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for the task: (returnval){ [ 793.680482] env[61824]: value = "task-1275454" [ 793.680482] env[61824]: _type = "Task" [ 793.680482] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.681858] env[61824]: DEBUG nova.network.neutron [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance_info_cache with network_info: [{"id": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "address": "fa:16:3e:ce:e9:8d", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape95afb5b-11", "ovs_interfaceid": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.721312] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fe403cc-8c76-4048-8991-9e3eb547aff2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.727848] env[61824]: DEBUG oslo_vmware.api [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.734570] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeaa507c-ca7c-44ed-91dc-dd77558ac9bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.753308] env[61824]: DEBUG nova.compute.provider_tree [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.839058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "68d46a6e-7903-4b5b-a5c2-928a97cbe23e" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.839842] env[61824]: DEBUG nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.951976] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275451, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.053442} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.952245] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.952964] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6eb7491-f925-4aed-9802-0ad424fc46f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.974267] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] f24881d3-a647-42f9-bf82-0065f3425a5f/f24881d3-a647-42f9-bf82-0065f3425a5f.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.977183] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cab6631-3544-4095-9bc3-dd49d7739217 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.996428] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275452, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442075} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.997588] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] e3e19a04-a29b-4473-833c-d8486703e9ad/e3e19a04-a29b-4473-833c-d8486703e9ad.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 793.997818] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.998128] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 793.998128] env[61824]: value = "task-1275455" [ 793.998128] env[61824]: _type = "Task" [ 793.998128] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.998336] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a1fc493-ed17-4c20-83a7-85edfe654e2c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.009274] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275455, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.010409] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 794.010409] env[61824]: value = "task-1275456" [ 794.010409] env[61824]: _type = "Task" [ 794.010409] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.019347] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275456, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.193512] env[61824]: DEBUG oslo_vmware.api [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.220734] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.220734] env[61824]: DEBUG nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Instance network_info: |[{"id": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "address": "fa:16:3e:ce:e9:8d", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape95afb5b-11", "ovs_interfaceid": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 794.221074] env[61824]: DEBUG oslo_concurrency.lockutils [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] Acquired lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.221157] env[61824]: DEBUG nova.network.neutron [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Refreshing network info cache for port e95afb5b-11f8-4ed4-abb0-a12b48c9c14d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 794.222389] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:e9:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd646f9d5-d2ad-4c22-bea5-85a965334de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e95afb5b-11f8-4ed4-abb0-a12b48c9c14d', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.232335] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Creating folder: Project (e51dd2f578344aa8afa5d6beccb54f1e). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.233695] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-767514b0-6473-478a-b0c5-6e8b45070d8e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.244320] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Created folder: Project (e51dd2f578344aa8afa5d6beccb54f1e) in parent group-v274074. [ 794.244521] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Creating folder: Instances. Parent ref: group-v274117. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 794.245132] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-135a2c5a-3fa0-4bdd-aa8e-79b30f9475c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.256753] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Created folder: Instances in parent group-v274117. [ 794.256753] env[61824]: DEBUG oslo.service.loopingcall [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.256753] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.257422] env[61824]: DEBUG nova.scheduler.client.report [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 794.260777] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f03b43d3-d4ae-4e03-9ef5-c3a596857240 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.288498] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.288498] env[61824]: value = "task-1275459" [ 794.288498] env[61824]: _type = "Task" [ 794.288498] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.296705] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275459, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.344659] env[61824]: DEBUG nova.compute.utils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.348233] env[61824]: DEBUG nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.348479] env[61824]: DEBUG nova.network.neutron [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 794.390763] env[61824]: DEBUG nova.policy [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '537045c0b3454ba2bb6254b024ffdce2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63dbbb82749c44c79138f5efb2cf043b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.511308] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275455, 'name': ReconfigVM_Task, 'duration_secs': 0.268455} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.514866] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Reconfigured VM instance instance-00000039 to attach disk [datastore1] f24881d3-a647-42f9-bf82-0065f3425a5f/f24881d3-a647-42f9-bf82-0065f3425a5f.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.515573] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a953fab8-bf6d-4daf-9ab0-93797c69e9fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.522378] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275456, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066649} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.523955] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 794.524553] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 794.524553] env[61824]: value = "task-1275460" [ 794.524553] env[61824]: _type = "Task" [ 794.524553] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.525321] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41490912-d4a6-46c7-928d-ce16fc427c99 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.552819] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] e3e19a04-a29b-4473-833c-d8486703e9ad/e3e19a04-a29b-4473-833c-d8486703e9ad.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 794.556796] env[61824]: DEBUG nova.network.neutron [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updated VIF entry in instance network info cache for port e95afb5b-11f8-4ed4-abb0-a12b48c9c14d. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 794.557162] env[61824]: DEBUG nova.network.neutron [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance_info_cache with network_info: [{"id": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "address": "fa:16:3e:ce:e9:8d", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape95afb5b-11", "ovs_interfaceid": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.558557] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37c07e7f-d024-4a83-843e-5371c5a5d5e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.573684] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275460, 'name': Rename_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.574921] env[61824]: DEBUG oslo_concurrency.lockutils [req-af826468-874d-4a32-8096-f5969140f64c req-f3d6a9df-59b8-48e1-a0fd-055bf873dcb6 service nova] Releasing lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.583964] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 794.583964] env[61824]: value = "task-1275461" [ 794.583964] env[61824]: _type = "Task" [ 794.583964] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.594919] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275461, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.693515] env[61824]: DEBUG oslo_vmware.api [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Task: {'id': task-1275454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.715135} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.693788] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.693973] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.694178] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.694352] env[61824]: INFO nova.compute.manager [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Took 1.65 seconds to destroy the instance on the hypervisor. [ 794.694629] env[61824]: DEBUG oslo.service.loopingcall [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.694843] env[61824]: DEBUG nova.compute.manager [-] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.694937] env[61824]: DEBUG nova.network.neutron [-] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 794.782417] env[61824]: DEBUG nova.network.neutron [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Successfully created port: 2947ff73-8e81-4e5a-9074-1db5399885ed {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.787159] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.792865] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.500s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.795926] env[61824]: INFO nova.compute.claims [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.811695] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275459, 'name': CreateVM_Task, 'duration_secs': 0.321719} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.811872] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 794.812867] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.813035] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.813442] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 794.813722] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9a3d31a-d8c4-4d31-bf2a-1a79c19606f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.818676] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 794.818676] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5216a3ce-32e4-4fbe-a009-9ee920894c5b" [ 794.818676] env[61824]: _type = "Task" [ 794.818676] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.829952] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5216a3ce-32e4-4fbe-a009-9ee920894c5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.851292] env[61824]: DEBUG nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.859598] env[61824]: INFO nova.scheduler.client.report [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleted allocations for instance 9fd8e236-3556-4b95-952f-3c324b896e29 [ 795.037691] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275460, 'name': Rename_Task, 'duration_secs': 0.156608} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.037952] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.038221] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5437c9ff-4747-481e-82d5-c836b0962897 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.043874] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 795.043874] env[61824]: value = "task-1275462" [ 795.043874] env[61824]: _type = "Task" [ 795.043874] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.051350] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275462, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.093660] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275461, 'name': ReconfigVM_Task, 'duration_secs': 0.251496} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.093932] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Reconfigured VM instance instance-0000003a to attach disk [datastore1] e3e19a04-a29b-4473-833c-d8486703e9ad/e3e19a04-a29b-4473-833c-d8486703e9ad.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.094528] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d4f6658-3541-4b08-89b9-04ec45942119 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.100506] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 795.100506] env[61824]: value = "task-1275463" [ 795.100506] env[61824]: _type = "Task" [ 795.100506] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.111899] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275463, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.332045] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5216a3ce-32e4-4fbe-a009-9ee920894c5b, 'name': SearchDatastore_Task, 'duration_secs': 0.008868} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.332045] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.332226] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.332412] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.332571] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.332719] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.332979] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37f5ad1b-81ae-497d-bb51-f966aa32f894 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.341567] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.341745] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 795.342595] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edb42189-a570-4f07-859c-c18d7429dbf1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.348264] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 795.348264] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a50012-f02e-36cb-6226-50cfe29b612e" [ 795.348264] env[61824]: _type = "Task" [ 795.348264] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.358877] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a50012-f02e-36cb-6226-50cfe29b612e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.367033] env[61824]: DEBUG oslo_concurrency.lockutils [None req-62ebebc9-e2a8-4a26-940d-62fcaa10d60c tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "9fd8e236-3556-4b95-952f-3c324b896e29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.835s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.434925] env[61824]: DEBUG nova.compute.manager [req-5acb1439-4a7d-41d8-8109-ad325348645a req-6a25a207-47ac-41f2-88e0-05a381d53a70 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Received event network-vif-deleted-cc85d94a-265d-4d8a-ade8-7ef62dfd117a {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.435225] env[61824]: INFO nova.compute.manager [req-5acb1439-4a7d-41d8-8109-ad325348645a req-6a25a207-47ac-41f2-88e0-05a381d53a70 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Neutron deleted interface cc85d94a-265d-4d8a-ade8-7ef62dfd117a; detaching it from the instance and deleting it from the info cache [ 795.435494] env[61824]: DEBUG nova.network.neutron [req-5acb1439-4a7d-41d8-8109-ad325348645a req-6a25a207-47ac-41f2-88e0-05a381d53a70 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.512875] env[61824]: DEBUG nova.network.neutron [-] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.557402] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275462, 'name': PowerOnVM_Task, 'duration_secs': 0.49319} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.557742] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.558032] env[61824]: INFO nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Took 8.85 seconds to spawn the instance on the hypervisor. [ 795.558298] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.559337] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71daa17c-d47c-452c-9f65-abcf2cdab9f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.610379] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275463, 'name': Rename_Task, 'duration_secs': 0.138429} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.610602] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 795.610829] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adc89956-f24d-4c29-8bba-32a0def956e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.616759] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 795.616759] env[61824]: value = "task-1275464" [ 795.616759] env[61824]: _type = "Task" [ 795.616759] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.623784] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275464, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.809458] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.809458] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.810056] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.810056] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.810056] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.812021] env[61824]: INFO nova.compute.manager [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Terminating instance [ 795.813751] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "refresh_cache-3f89fb2d-99c1-4cf7-83aa-db033211a2ce" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.813902] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquired lock "refresh_cache-3f89fb2d-99c1-4cf7-83aa-db033211a2ce" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.814078] env[61824]: DEBUG nova.network.neutron [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 795.861283] env[61824]: DEBUG nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.863117] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a50012-f02e-36cb-6226-50cfe29b612e, 'name': SearchDatastore_Task, 'duration_secs': 0.008058} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.866051] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c43d1659-fc01-4677-8aed-7f3f1f983e90 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.871911] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 795.871911] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52ed60fd-4dad-fc54-0642-e8434a569b43" [ 795.871911] env[61824]: _type = "Task" [ 795.871911] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.879755] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ed60fd-4dad-fc54-0642-e8434a569b43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.887914] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.888158] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.888312] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.888493] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.888633] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.888779] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.888981] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.889319] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.889319] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.889467] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.889685] env[61824]: DEBUG nova.virt.hardware [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.890805] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cc49b5-6e86-4186-bb13-6f912eb672a9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.898836] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622934cd-ea5d-4769-a945-429aad63054d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.938453] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03015967-77df-4358-8efd-7155a5c948ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.947050] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ed17f6-354b-4216-851c-7499a875c34c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.974704] env[61824]: DEBUG nova.compute.manager [req-5acb1439-4a7d-41d8-8109-ad325348645a req-6a25a207-47ac-41f2-88e0-05a381d53a70 service nova] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Detach interface failed, port_id=cc85d94a-265d-4d8a-ade8-7ef62dfd117a, reason: Instance cc13d6a5-6b99-40bf-8978-09d284b4ed17 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 796.015645] env[61824]: INFO nova.compute.manager [-] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Took 1.32 seconds to deallocate network for instance. [ 796.081953] env[61824]: INFO nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Took 37.01 seconds to build instance. [ 796.126110] env[61824]: DEBUG oslo_vmware.api [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275464, 'name': PowerOnVM_Task, 'duration_secs': 0.469937} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.128518] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 796.128752] env[61824]: INFO nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Took 7.06 seconds to spawn the instance on the hypervisor. [ 796.128928] env[61824]: DEBUG nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.129831] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f10d19-4ff1-4a95-8bea-f66bb2aede7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.149856] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b39c0dcd-b895-471c-b5d1-273419a55300 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.157796] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335a2a31-0f09-4b96-b5fd-388aab7870e4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.188079] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833d39f7-b69e-4b9e-8862-d0dbbf5b8bed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.196392] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daaeefa5-7fdd-44e3-8e54-033266ff143c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.211442] env[61824]: DEBUG nova.compute.provider_tree [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.335019] env[61824]: DEBUG nova.network.neutron [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 796.362487] env[61824]: DEBUG nova.compute.manager [req-a52414ac-3ba9-4581-8802-b0f39e09545d req-81afc197-4491-4098-917f-79d7089b860d service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Received event network-vif-plugged-2947ff73-8e81-4e5a-9074-1db5399885ed {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.362713] env[61824]: DEBUG oslo_concurrency.lockutils [req-a52414ac-3ba9-4581-8802-b0f39e09545d req-81afc197-4491-4098-917f-79d7089b860d service nova] Acquiring lock "8522fe80-4d6d-4474-ae63-00daa0dc7349-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.362913] env[61824]: DEBUG oslo_concurrency.lockutils [req-a52414ac-3ba9-4581-8802-b0f39e09545d req-81afc197-4491-4098-917f-79d7089b860d service nova] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.364318] env[61824]: DEBUG oslo_concurrency.lockutils [req-a52414ac-3ba9-4581-8802-b0f39e09545d req-81afc197-4491-4098-917f-79d7089b860d service nova] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.364318] env[61824]: DEBUG nova.compute.manager [req-a52414ac-3ba9-4581-8802-b0f39e09545d req-81afc197-4491-4098-917f-79d7089b860d service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] No waiting events found dispatching network-vif-plugged-2947ff73-8e81-4e5a-9074-1db5399885ed {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.364318] env[61824]: WARNING nova.compute.manager [req-a52414ac-3ba9-4581-8802-b0f39e09545d req-81afc197-4491-4098-917f-79d7089b860d service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Received unexpected event network-vif-plugged-2947ff73-8e81-4e5a-9074-1db5399885ed for instance with vm_state building and task_state spawning. [ 796.383142] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ed60fd-4dad-fc54-0642-e8434a569b43, 'name': SearchDatastore_Task, 'duration_secs': 0.008645} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.383486] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.383797] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] a33e2573-6d2f-4d07-89d4-f1c6d9d9b382/a33e2573-6d2f-4d07-89d4-f1c6d9d9b382.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 796.384269] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c62a4ce-c524-48db-8b3d-7f14e1dd6dd7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.391108] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 796.391108] env[61824]: value = "task-1275465" [ 796.391108] env[61824]: _type = "Task" [ 796.391108] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.392116] env[61824]: DEBUG nova.network.neutron [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.403054] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275465, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.523115] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.583434] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.194s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.648604] env[61824]: INFO nova.compute.manager [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Took 35.75 seconds to build instance. [ 796.673617] env[61824]: DEBUG nova.network.neutron [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Successfully updated port: 2947ff73-8e81-4e5a-9074-1db5399885ed {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.714843] env[61824]: DEBUG nova.scheduler.client.report [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 796.901186] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Releasing lock "refresh_cache-3f89fb2d-99c1-4cf7-83aa-db033211a2ce" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.901625] env[61824]: DEBUG nova.compute.manager [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 796.901815] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 796.902118] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275465, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.902937] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7509118e-0f70-43dc-8a4c-1a53b460bad8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.910112] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 796.910465] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3b1f084-b07e-4e7e-90ce-c49ff6e4ec25 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.917329] env[61824]: DEBUG oslo_vmware.api [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 796.917329] env[61824]: value = "task-1275466" [ 796.917329] env[61824]: _type = "Task" [ 796.917329] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.927715] env[61824]: DEBUG oslo_vmware.api [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.086403] env[61824]: DEBUG nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.150956] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d3dfa044-976b-4a2e-8d61-b578a8ccee92 tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.739s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.175838] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "refresh_cache-8522fe80-4d6d-4474-ae63-00daa0dc7349" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.175998] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquired lock "refresh_cache-8522fe80-4d6d-4474-ae63-00daa0dc7349" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.176171] env[61824]: DEBUG nova.network.neutron [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 797.220698] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.220698] env[61824]: DEBUG nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.226294] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.301s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.229749] env[61824]: INFO nova.compute.claims [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.401495] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275465, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.627557} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.401736] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] a33e2573-6d2f-4d07-89d4-f1c6d9d9b382/a33e2573-6d2f-4d07-89d4-f1c6d9d9b382.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.401945] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.402274] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-72c414ef-6346-4fe4-992e-82f52acd9465 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.409316] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 797.409316] env[61824]: value = "task-1275467" [ 797.409316] env[61824]: _type = "Task" [ 797.409316] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.417293] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275467, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.426946] env[61824]: DEBUG oslo_vmware.api [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275466, 'name': PowerOffVM_Task, 'duration_secs': 0.132432} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.427291] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 797.427372] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 797.427636] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39e719ae-62db-42c1-bb73-7935b7aba478 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.451742] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 797.451996] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 797.452194] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleting the datastore file [datastore2] 3f89fb2d-99c1-4cf7-83aa-db033211a2ce {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.452945] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d5607a6-5e80-4855-afc0-31775c9f3ea9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.459315] env[61824]: DEBUG oslo_vmware.api [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for the task: (returnval){ [ 797.459315] env[61824]: value = "task-1275469" [ 797.459315] env[61824]: _type = "Task" [ 797.459315] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.468794] env[61824]: DEBUG oslo_vmware.api [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275469, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.608347] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.652928] env[61824]: DEBUG nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.716829] env[61824]: DEBUG nova.network.neutron [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.733791] env[61824]: DEBUG nova.compute.utils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 797.736869] env[61824]: DEBUG nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 797.737060] env[61824]: DEBUG nova.network.neutron [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 797.774951] env[61824]: DEBUG nova.policy [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb8bfb13b7bb42a6bc03dedb4a38599b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'af46b7fc8900414c91c9f64d41ebe314', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.865386] env[61824]: DEBUG nova.network.neutron [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Updating instance_info_cache with network_info: [{"id": "2947ff73-8e81-4e5a-9074-1db5399885ed", "address": "fa:16:3e:6a:db:9a", "network": {"id": "aef63786-04fc-4804-b18c-f60ed952ae16", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-806971963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63dbbb82749c44c79138f5efb2cf043b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0549820d-5649-40bc-ad6e-9ae27b384d90", "external-id": "nsx-vlan-transportzone-434", "segmentation_id": 434, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2947ff73-8e", "ovs_interfaceid": "2947ff73-8e81-4e5a-9074-1db5399885ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.919920] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275467, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.249294} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.920205] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.921081] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd2af3c-494e-4d01-834e-1f05844fdebd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.947662] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] a33e2573-6d2f-4d07-89d4-f1c6d9d9b382/a33e2573-6d2f-4d07-89d4-f1c6d9d9b382.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.947662] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "f24881d3-a647-42f9-bf82-0065f3425a5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.947876] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.947911] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "f24881d3-a647-42f9-bf82-0065f3425a5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.948105] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.948662] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.950164] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c81ca55a-cb5f-4f7e-9e7c-0d27d836dc44 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.964475] env[61824]: INFO nova.compute.manager [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Terminating instance [ 797.969586] env[61824]: DEBUG nova.compute.manager [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 797.969792] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.970482] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757b2d7d-43ac-4979-bafd-ac660c9446ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.974910] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 797.974910] env[61824]: value = "task-1275470" [ 797.974910] env[61824]: _type = "Task" [ 797.974910] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.981695] env[61824]: DEBUG oslo_vmware.api [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Task: {'id': task-1275469, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093008} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.983433] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.983623] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 797.983801] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 797.983973] env[61824]: INFO nova.compute.manager [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Took 1.08 seconds to destroy the instance on the hypervisor. [ 797.984227] env[61824]: DEBUG oslo.service.loopingcall [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.985036] env[61824]: DEBUG nova.compute.manager [-] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 797.985143] env[61824]: DEBUG nova.network.neutron [-] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.991135] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 797.991382] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275470, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.991860] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86bd9427-2f52-43ef-baaf-83d8dd238978 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.998149] env[61824]: DEBUG oslo_vmware.api [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 797.998149] env[61824]: value = "task-1275471" [ 797.998149] env[61824]: _type = "Task" [ 797.998149] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.005757] env[61824]: DEBUG oslo_vmware.api [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.009084] env[61824]: DEBUG nova.network.neutron [-] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 798.012117] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "e3e19a04-a29b-4473-833c-d8486703e9ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.012338] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.012531] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "e3e19a04-a29b-4473-833c-d8486703e9ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.012719] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.012883] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.014913] env[61824]: INFO nova.compute.manager [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Terminating instance [ 798.020021] env[61824]: DEBUG nova.compute.manager [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 798.020213] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 798.020998] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23eedc5f-2442-4d96-9376-da33fd09fe31 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.028631] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 798.028914] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3fb87562-73c1-4870-a2ef-faa7ac2ba0d2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.037916] env[61824]: DEBUG oslo_vmware.api [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 798.037916] env[61824]: value = "task-1275472" [ 798.037916] env[61824]: _type = "Task" [ 798.037916] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.045658] env[61824]: DEBUG oslo_vmware.api [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.065925] env[61824]: DEBUG nova.network.neutron [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Successfully created port: 64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.176469] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.243935] env[61824]: DEBUG nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 798.367917] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Releasing lock "refresh_cache-8522fe80-4d6d-4474-ae63-00daa0dc7349" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.368248] env[61824]: DEBUG nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Instance network_info: |[{"id": "2947ff73-8e81-4e5a-9074-1db5399885ed", "address": "fa:16:3e:6a:db:9a", "network": {"id": "aef63786-04fc-4804-b18c-f60ed952ae16", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-806971963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63dbbb82749c44c79138f5efb2cf043b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0549820d-5649-40bc-ad6e-9ae27b384d90", "external-id": "nsx-vlan-transportzone-434", "segmentation_id": 434, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2947ff73-8e", "ovs_interfaceid": "2947ff73-8e81-4e5a-9074-1db5399885ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.368680] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:db:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0549820d-5649-40bc-ad6e-9ae27b384d90', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2947ff73-8e81-4e5a-9074-1db5399885ed', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.376386] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Creating folder: Project (63dbbb82749c44c79138f5efb2cf043b). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.379125] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e731aa2-e53f-442f-9891-7ecac93133e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.390042] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Created folder: Project (63dbbb82749c44c79138f5efb2cf043b) in parent group-v274074. [ 798.390241] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Creating folder: Instances. Parent ref: group-v274120. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.392645] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7a9639a-3816-4936-b234-7411c618e6c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.401158] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Created folder: Instances in parent group-v274120. [ 798.401387] env[61824]: DEBUG oslo.service.loopingcall [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.401583] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 798.401772] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c1b7d8b-b420-4cd5-b1e9-df1bf8ffd72d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.423338] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.423338] env[61824]: value = "task-1275475" [ 798.423338] env[61824]: _type = "Task" [ 798.423338] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.433247] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275475, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.484253] env[61824]: DEBUG nova.compute.manager [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Received event network-changed-2947ff73-8e81-4e5a-9074-1db5399885ed {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.484471] env[61824]: DEBUG nova.compute.manager [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Refreshing instance network info cache due to event network-changed-2947ff73-8e81-4e5a-9074-1db5399885ed. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 798.484685] env[61824]: DEBUG oslo_concurrency.lockutils [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] Acquiring lock "refresh_cache-8522fe80-4d6d-4474-ae63-00daa0dc7349" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.484822] env[61824]: DEBUG oslo_concurrency.lockutils [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] Acquired lock "refresh_cache-8522fe80-4d6d-4474-ae63-00daa0dc7349" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.485084] env[61824]: DEBUG nova.network.neutron [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Refreshing network info cache for port 2947ff73-8e81-4e5a-9074-1db5399885ed {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.494482] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275470, 'name': ReconfigVM_Task, 'duration_secs': 0.413364} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.495278] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Reconfigured VM instance instance-0000003b to attach disk [datastore1] a33e2573-6d2f-4d07-89d4-f1c6d9d9b382/a33e2573-6d2f-4d07-89d4-f1c6d9d9b382.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.495930] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc90cf9e-84ed-4363-9735-658f96c18892 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.506171] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 798.506171] env[61824]: value = "task-1275476" [ 798.506171] env[61824]: _type = "Task" [ 798.506171] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.512738] env[61824]: DEBUG nova.network.neutron [-] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.514118] env[61824]: DEBUG oslo_vmware.api [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275471, 'name': PowerOffVM_Task, 'duration_secs': 0.21362} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.514832] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 798.515011] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 798.515248] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f754edc-aae3-4e84-86a8-14cd0975ab0b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.522054] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275476, 'name': Rename_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.529776] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144b3bcb-41c1-4a5b-b889-0d29aa42eb44 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.542666] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ed7948-e3ec-4226-9e4a-698262d583ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.552321] env[61824]: DEBUG oslo_vmware.api [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275472, 'name': PowerOffVM_Task, 'duration_secs': 0.205392} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.579237] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 798.579472] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 798.583914] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ed7bdf1-0891-4e9d-ab5d-f08e83efd5df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.583914] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6e8ecb-fee7-4936-a949-58550c71b242 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.586478] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 798.586666] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 798.586832] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleting the datastore file [datastore1] f24881d3-a647-42f9-bf82-0065f3425a5f {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.587234] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04b0b45c-654b-4b18-a3e5-5847c31faf4d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.596035] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f1c326-57a5-4ca8-aa67-c614e79d3650 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.599911] env[61824]: DEBUG oslo_vmware.api [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 798.599911] env[61824]: value = "task-1275479" [ 798.599911] env[61824]: _type = "Task" [ 798.599911] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.611522] env[61824]: DEBUG nova.compute.provider_tree [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.617461] env[61824]: DEBUG oslo_vmware.api [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.651997] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 798.652271] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 798.652454] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleting the datastore file [datastore1] e3e19a04-a29b-4473-833c-d8486703e9ad {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.652749] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b7e467f-4e8e-4986-8cf4-4317cedf4569 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.659319] env[61824]: DEBUG oslo_vmware.api [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for the task: (returnval){ [ 798.659319] env[61824]: value = "task-1275480" [ 798.659319] env[61824]: _type = "Task" [ 798.659319] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.667953] env[61824]: DEBUG oslo_vmware.api [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275480, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.749725] env[61824]: INFO nova.virt.block_device [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Booting with volume 32104921-3bcd-4691-86e7-9c3c0a08ed72 at /dev/sda [ 798.790025] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a435f5c0-9178-40df-a8d1-6ccd3785b01c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.799615] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7983ff0f-54c9-44dc-9685-f0e67b7a7cf9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.823307] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c10925e1-ad71-4090-95eb-5bbe1c0576d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.830592] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912c9976-893e-4e34-911f-4fbc3ce644fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.853122] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d986fd-0e2a-4f28-872b-46085eeb010d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.859013] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aea397f-269b-434a-8f57-309995692432 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.874813] env[61824]: DEBUG nova.virt.block_device [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Updating existing volume attachment record: fbfbb2e1-9258-4818-923c-7dcb283efe3e {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 798.934951] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275475, 'name': CreateVM_Task, 'duration_secs': 0.318613} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.935155] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.935883] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.936055] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.936377] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.936635] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4ddec25-4688-47e1-8d41-fb23f3909f12 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.941171] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 798.941171] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c439bc-de3f-163a-53f0-c60ef9db188b" [ 798.941171] env[61824]: _type = "Task" [ 798.941171] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.950071] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c439bc-de3f-163a-53f0-c60ef9db188b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.015748] env[61824]: INFO nova.compute.manager [-] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Took 1.03 seconds to deallocate network for instance. [ 799.016125] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275476, 'name': Rename_Task, 'duration_secs': 0.163242} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.017998] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.018268] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-840a3076-4efa-4d04-969b-1ef1840a82c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.029043] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 799.029043] env[61824]: value = "task-1275481" [ 799.029043] env[61824]: _type = "Task" [ 799.029043] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.041948] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275481, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.108947] env[61824]: DEBUG oslo_vmware.api [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162021} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.109201] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.109449] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 799.109653] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.109820] env[61824]: INFO nova.compute.manager [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 799.110063] env[61824]: DEBUG oslo.service.loopingcall [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.110250] env[61824]: DEBUG nova.compute.manager [-] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.110340] env[61824]: DEBUG nova.network.neutron [-] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 799.114671] env[61824]: DEBUG nova.scheduler.client.report [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.169571] env[61824]: DEBUG oslo_vmware.api [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Task: {'id': task-1275480, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148279} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.169823] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.169998] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 799.170184] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.170349] env[61824]: INFO nova.compute.manager [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Took 1.15 seconds to destroy the instance on the hypervisor. [ 799.170570] env[61824]: DEBUG oslo.service.loopingcall [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.170746] env[61824]: DEBUG nova.compute.manager [-] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.170831] env[61824]: DEBUG nova.network.neutron [-] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 799.428288] env[61824]: DEBUG nova.network.neutron [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Updated VIF entry in instance network info cache for port 2947ff73-8e81-4e5a-9074-1db5399885ed. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.428726] env[61824]: DEBUG nova.network.neutron [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Updating instance_info_cache with network_info: [{"id": "2947ff73-8e81-4e5a-9074-1db5399885ed", "address": "fa:16:3e:6a:db:9a", "network": {"id": "aef63786-04fc-4804-b18c-f60ed952ae16", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-806971963-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "63dbbb82749c44c79138f5efb2cf043b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0549820d-5649-40bc-ad6e-9ae27b384d90", "external-id": "nsx-vlan-transportzone-434", "segmentation_id": 434, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2947ff73-8e", "ovs_interfaceid": "2947ff73-8e81-4e5a-9074-1db5399885ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.457315] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c439bc-de3f-163a-53f0-c60ef9db188b, 'name': SearchDatastore_Task, 'duration_secs': 0.009652} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.457850] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.458221] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.458765] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.459099] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.459559] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.460274] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-931e2a24-6b99-4fc8-8c31-cbb2a9a354d6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.469863] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.470298] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.474746] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82f459a3-3bb2-4fa9-93a5-deeb030693c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.483412] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 799.483412] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52598256-36fd-8054-1875-030726a5be80" [ 799.483412] env[61824]: _type = "Task" [ 799.483412] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.490429] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52598256-36fd-8054-1875-030726a5be80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.526688] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.538143] env[61824]: DEBUG oslo_vmware.api [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275481, 'name': PowerOnVM_Task, 'duration_secs': 0.448315} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.538396] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 799.538618] env[61824]: INFO nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Took 8.09 seconds to spawn the instance on the hypervisor. [ 799.538799] env[61824]: DEBUG nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.541889] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2813c5eb-4c6e-4d52-a0c4-a7a954ee947f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.570113] env[61824]: DEBUG nova.compute.manager [req-a2b09424-f455-40d6-a6fd-40e60dc0d3e6 req-dd1e2606-e6b8-4768-806d-7ec6a326a557 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Received event network-vif-plugged-64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.570339] env[61824]: DEBUG oslo_concurrency.lockutils [req-a2b09424-f455-40d6-a6fd-40e60dc0d3e6 req-dd1e2606-e6b8-4768-806d-7ec6a326a557 service nova] Acquiring lock "6235a9b9-19c0-4917-938a-bca82a2ca05e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.572024] env[61824]: DEBUG oslo_concurrency.lockutils [req-a2b09424-f455-40d6-a6fd-40e60dc0d3e6 req-dd1e2606-e6b8-4768-806d-7ec6a326a557 service nova] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.572024] env[61824]: DEBUG oslo_concurrency.lockutils [req-a2b09424-f455-40d6-a6fd-40e60dc0d3e6 req-dd1e2606-e6b8-4768-806d-7ec6a326a557 service nova] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.572024] env[61824]: DEBUG nova.compute.manager [req-a2b09424-f455-40d6-a6fd-40e60dc0d3e6 req-dd1e2606-e6b8-4768-806d-7ec6a326a557 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] No waiting events found dispatching network-vif-plugged-64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.572550] env[61824]: WARNING nova.compute.manager [req-a2b09424-f455-40d6-a6fd-40e60dc0d3e6 req-dd1e2606-e6b8-4768-806d-7ec6a326a557 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Received unexpected event network-vif-plugged-64000f4d-cb2c-4d63-95b7-97218e98723c for instance with vm_state building and task_state block_device_mapping. [ 799.599147] env[61824]: DEBUG nova.network.neutron [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Successfully updated port: 64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.619250] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.619789] env[61824]: DEBUG nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 799.623106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.173s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.628383] env[61824]: INFO nova.compute.claims [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.934925] env[61824]: DEBUG oslo_concurrency.lockutils [req-0224bb84-deb2-4bda-827f-51835ca144eb req-b2b92827-d94e-4493-a84d-f323fabebb84 service nova] Releasing lock "refresh_cache-8522fe80-4d6d-4474-ae63-00daa0dc7349" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.993797] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52598256-36fd-8054-1875-030726a5be80, 'name': SearchDatastore_Task, 'duration_secs': 0.009842} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.994569] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0d2ad07-839e-49f3-ad82-4883243de000 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.000264] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 800.000264] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]520102a2-ee71-480f-567e-67597e9ab415" [ 800.000264] env[61824]: _type = "Task" [ 800.000264] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.008205] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520102a2-ee71-480f-567e-67597e9ab415, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.061911] env[61824]: INFO nova.compute.manager [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Took 37.35 seconds to build instance. [ 800.101511] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Acquiring lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.101668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Acquired lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.101818] env[61824]: DEBUG nova.network.neutron [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.134026] env[61824]: DEBUG nova.compute.utils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.136441] env[61824]: DEBUG nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.136441] env[61824]: DEBUG nova.network.neutron [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 800.182135] env[61824]: DEBUG nova.policy [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '11968f526d0a4959a49d80a3fe90eb18', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70e6557f001f486b9c0d7923bd161981', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 800.212367] env[61824]: DEBUG nova.network.neutron [-] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.218037] env[61824]: DEBUG nova.network.neutron [-] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.510501] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520102a2-ee71-480f-567e-67597e9ab415, 'name': SearchDatastore_Task, 'duration_secs': 0.009645} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.510753] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.511018] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 8522fe80-4d6d-4474-ae63-00daa0dc7349/8522fe80-4d6d-4474-ae63-00daa0dc7349.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.511268] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6433e1ee-1f88-4afc-8564-dfd39e9d25a8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.516198] env[61824]: DEBUG nova.compute.manager [req-5b2061dd-5195-4459-9781-cc8fb5e3d0d9 req-5bc66eb6-f35b-4b9b-a73f-fcf9c6bb7265 service nova] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Received event network-vif-deleted-47dcce78-bb07-4111-9322-e92f8f6763b1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.516276] env[61824]: DEBUG nova.compute.manager [req-5b2061dd-5195-4459-9781-cc8fb5e3d0d9 req-5bc66eb6-f35b-4b9b-a73f-fcf9c6bb7265 service nova] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Received event network-vif-deleted-61a97b86-36a1-41d7-b3af-38ba07db954d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.522960] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 800.522960] env[61824]: value = "task-1275482" [ 800.522960] env[61824]: _type = "Task" [ 800.522960] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.531630] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275482, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.550418] env[61824]: DEBUG nova.network.neutron [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Successfully created port: b9d880b8-2784-4570-9726-b6f545278cf2 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.564164] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6a4653a3-7c07-469a-8e8f-fc5705b3976b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.625s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.637206] env[61824]: DEBUG nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 800.651312] env[61824]: DEBUG nova.network.neutron [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.714996] env[61824]: INFO nova.compute.manager [-] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Took 1.54 seconds to deallocate network for instance. [ 800.723019] env[61824]: INFO nova.compute.manager [-] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Took 1.61 seconds to deallocate network for instance. [ 800.934561] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1502ca-8ae3-4a0f-88dc-0fd43ef45df4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.943279] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf066ce4-37b7-4041-9ecf-3e5db2f1d12a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.974956] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56afb45f-60f1-4ec3-9a08-e27bd57822d4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.982938] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ce631a-ccf3-4821-a7f6-91f18ab3f1c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.999186] env[61824]: DEBUG nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.999766] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.999977] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.000143] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.000322] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.000464] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.000608] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.000825] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.000954] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.002131] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.002340] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.002525] env[61824]: DEBUG nova.virt.hardware [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.003060] env[61824]: DEBUG nova.compute.provider_tree [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.007616] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884c419e-6888-4eaf-a9af-f47fe4766362 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.015318] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a168dd1f-987e-4b5d-98b0-f6b2a0daf5ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.038933] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275482, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471827} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.039132] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 8522fe80-4d6d-4474-ae63-00daa0dc7349/8522fe80-4d6d-4474-ae63-00daa0dc7349.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.039352] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.039995] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-91698475-ffb3-4f31-b33b-1e5f7fa465f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.045806] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 801.045806] env[61824]: value = "task-1275483" [ 801.045806] env[61824]: _type = "Task" [ 801.045806] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.053354] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275483, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.066341] env[61824]: DEBUG nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 801.105372] env[61824]: DEBUG nova.network.neutron [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Updating instance_info_cache with network_info: [{"id": "64000f4d-cb2c-4d63-95b7-97218e98723c", "address": "fa:16:3e:87:33:70", "network": {"id": "33163902-81f7-4ebc-97ee-1d81884d3375", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-712910259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af46b7fc8900414c91c9f64d41ebe314", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64000f4d-cb", "ovs_interfaceid": "64000f4d-cb2c-4d63-95b7-97218e98723c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.224982] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.229935] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.512730] env[61824]: DEBUG nova.scheduler.client.report [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 801.556028] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275483, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102261} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.556028] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.556598] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8482654f-b1f0-48c7-8fc9-0050281cb5f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.578882] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 8522fe80-4d6d-4474-ae63-00daa0dc7349/8522fe80-4d6d-4474-ae63-00daa0dc7349.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.581377] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b5cce7a-48c8-46bc-a5a3-e02815cf0845 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.601117] env[61824]: DEBUG nova.compute.manager [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Received event network-changed-64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.601320] env[61824]: DEBUG nova.compute.manager [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Refreshing instance network info cache due to event network-changed-64000f4d-cb2c-4d63-95b7-97218e98723c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 801.601512] env[61824]: DEBUG oslo_concurrency.lockutils [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] Acquiring lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.606092] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 801.606092] env[61824]: value = "task-1275484" [ 801.606092] env[61824]: _type = "Task" [ 801.606092] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.611370] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.611370] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Releasing lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.611488] env[61824]: DEBUG nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Instance network_info: |[{"id": "64000f4d-cb2c-4d63-95b7-97218e98723c", "address": "fa:16:3e:87:33:70", "network": {"id": "33163902-81f7-4ebc-97ee-1d81884d3375", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-712910259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af46b7fc8900414c91c9f64d41ebe314", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64000f4d-cb", "ovs_interfaceid": "64000f4d-cb2c-4d63-95b7-97218e98723c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 801.611887] env[61824]: DEBUG oslo_concurrency.lockutils [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] Acquired lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.612068] env[61824]: DEBUG nova.network.neutron [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Refreshing network info cache for port 64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 801.613309] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:33:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7b4bfde-f109-4f64-adab-e7f06b80685d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '64000f4d-cb2c-4d63-95b7-97218e98723c', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.620956] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Creating folder: Project (af46b7fc8900414c91c9f64d41ebe314). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.627616] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9de50974-da0a-4f2f-91cf-6f9bf62c8a56 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.629638] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275484, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.640951] env[61824]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 801.641087] env[61824]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61824) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 801.641492] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Folder already exists: Project (af46b7fc8900414c91c9f64d41ebe314). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 801.641621] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Creating folder: Instances. Parent ref: group-v274090. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.641834] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e5b23ba6-6d63-4238-9dcc-dd8bdd924205 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.650458] env[61824]: DEBUG nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 801.653757] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Created folder: Instances in parent group-v274090. [ 801.653992] env[61824]: DEBUG oslo.service.loopingcall [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.654195] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 801.654399] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9fb686a3-b8d7-4d32-adc0-d1b79d405d17 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.679431] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.679431] env[61824]: value = "task-1275487" [ 801.679431] env[61824]: _type = "Task" [ 801.679431] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.681128] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.682528] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.682528] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.682528] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.682528] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.682528] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.682778] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.682778] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.682778] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.682778] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.682778] env[61824]: DEBUG nova.virt.hardware [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.683667] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2051c4-313f-4177-9e04-aeb623c2f739 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.695894] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275487, 'name': CreateVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.698219] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bb4c6f-c95c-49d3-92d5-33590fb08abe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.882674] env[61824]: DEBUG nova.network.neutron [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Updated VIF entry in instance network info cache for port 64000f4d-cb2c-4d63-95b7-97218e98723c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.883194] env[61824]: DEBUG nova.network.neutron [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Updating instance_info_cache with network_info: [{"id": "64000f4d-cb2c-4d63-95b7-97218e98723c", "address": "fa:16:3e:87:33:70", "network": {"id": "33163902-81f7-4ebc-97ee-1d81884d3375", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-712910259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af46b7fc8900414c91c9f64d41ebe314", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64000f4d-cb", "ovs_interfaceid": "64000f4d-cb2c-4d63-95b7-97218e98723c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.018470] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.018737] env[61824]: DEBUG nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.021740] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.047s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.023394] env[61824]: INFO nova.compute.claims [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.118906] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275484, 'name': ReconfigVM_Task, 'duration_secs': 0.282821} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.118906] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 8522fe80-4d6d-4474-ae63-00daa0dc7349/8522fe80-4d6d-4474-ae63-00daa0dc7349.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 802.118906] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-38404aff-8b94-49c1-947d-ac15d01dd3c7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.125021] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 802.125021] env[61824]: value = "task-1275488" [ 802.125021] env[61824]: _type = "Task" [ 802.125021] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.135950] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275488, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.193775] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275487, 'name': CreateVM_Task, 'duration_secs': 0.504154} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.193959] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.194680] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'boot_index': 0, 'guest_format': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274093', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'name': 'volume-32104921-3bcd-4691-86e7-9c3c0a08ed72', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6235a9b9-19c0-4917-938a-bca82a2ca05e', 'attached_at': '', 'detached_at': '', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'serial': '32104921-3bcd-4691-86e7-9c3c0a08ed72'}, 'mount_device': '/dev/sda', 'attachment_id': 'fbfbb2e1-9258-4818-923c-7dcb283efe3e', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 802.194912] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Root volume attach. Driver type: vmdk {{(pid=61824) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 802.195761] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2af66f5-945f-4e83-a006-590faa062ac1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.203836] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80321854-0110-4231-bafb-f0cf81446ed3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.210017] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0f8b00-1f4d-478f-9129-baebe1457b2e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.216192] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-1db8877b-86e7-4f2c-82b5-cf2549678c46 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.222897] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 802.222897] env[61824]: value = "task-1275489" [ 802.222897] env[61824]: _type = "Task" [ 802.222897] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.230500] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275489, 'name': RelocateVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.385408] env[61824]: DEBUG oslo_concurrency.lockutils [req-e0f7781b-87a0-4fb5-847a-5dff50021877 req-bcaf9837-0a94-40c1-9735-22f8b0f89cb7 service nova] Releasing lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.531847] env[61824]: DEBUG nova.compute.utils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.534436] env[61824]: DEBUG nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.534613] env[61824]: DEBUG nova.network.neutron [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 802.616815] env[61824]: DEBUG nova.compute.manager [req-b84e81d0-5692-44fa-95fe-7defac51be45 req-b00f723d-bebe-4cc9-b91e-31e5f3841b6e service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Received event network-vif-plugged-b9d880b8-2784-4570-9726-b6f545278cf2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.617040] env[61824]: DEBUG oslo_concurrency.lockutils [req-b84e81d0-5692-44fa-95fe-7defac51be45 req-b00f723d-bebe-4cc9-b91e-31e5f3841b6e service nova] Acquiring lock "0201e783-ed75-4cee-8df4-efac97b4ca19-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.617258] env[61824]: DEBUG oslo_concurrency.lockutils [req-b84e81d0-5692-44fa-95fe-7defac51be45 req-b00f723d-bebe-4cc9-b91e-31e5f3841b6e service nova] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.617474] env[61824]: DEBUG oslo_concurrency.lockutils [req-b84e81d0-5692-44fa-95fe-7defac51be45 req-b00f723d-bebe-4cc9-b91e-31e5f3841b6e service nova] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.617657] env[61824]: DEBUG nova.compute.manager [req-b84e81d0-5692-44fa-95fe-7defac51be45 req-b00f723d-bebe-4cc9-b91e-31e5f3841b6e service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] No waiting events found dispatching network-vif-plugged-b9d880b8-2784-4570-9726-b6f545278cf2 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.617820] env[61824]: WARNING nova.compute.manager [req-b84e81d0-5692-44fa-95fe-7defac51be45 req-b00f723d-bebe-4cc9-b91e-31e5f3841b6e service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Received unexpected event network-vif-plugged-b9d880b8-2784-4570-9726-b6f545278cf2 for instance with vm_state building and task_state spawning. [ 802.624179] env[61824]: DEBUG nova.policy [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a975aca3353469a948c52b035e17fcc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '502b41859ba843838d2188c82f70ae08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.635301] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275488, 'name': Rename_Task, 'duration_secs': 0.148051} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.635596] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.635833] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac7dc3d1-7d5e-4826-8031-3161287cf336 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.642017] env[61824]: DEBUG nova.network.neutron [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Successfully updated port: b9d880b8-2784-4570-9726-b6f545278cf2 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 802.644174] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 802.644174] env[61824]: value = "task-1275490" [ 802.644174] env[61824]: _type = "Task" [ 802.644174] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.652785] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.733528] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275489, 'name': RelocateVM_Task} progress is 19%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.874757] env[61824]: DEBUG nova.compute.manager [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Stashing vm_state: active {{(pid=61824) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 803.041778] env[61824]: DEBUG nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.148996] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "refresh_cache-0201e783-ed75-4cee-8df4-efac97b4ca19" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.149164] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquired lock "refresh_cache-0201e783-ed75-4cee-8df4-efac97b4ca19" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.149312] env[61824]: DEBUG nova.network.neutron [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.155916] env[61824]: DEBUG oslo_vmware.api [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275490, 'name': PowerOnVM_Task, 'duration_secs': 0.447808} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.158894] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.159107] env[61824]: INFO nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Took 7.30 seconds to spawn the instance on the hypervisor. [ 803.159288] env[61824]: DEBUG nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.160780] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1613adc9-9bd7-459f-b4d9-993d6c0d74ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.237788] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275489, 'name': RelocateVM_Task, 'duration_secs': 0.710691} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.238399] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 803.238639] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274093', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'name': 'volume-32104921-3bcd-4691-86e7-9c3c0a08ed72', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6235a9b9-19c0-4917-938a-bca82a2ca05e', 'attached_at': '', 'detached_at': '', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'serial': '32104921-3bcd-4691-86e7-9c3c0a08ed72'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 803.239420] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7713f1d3-1aa7-4fcc-ae16-31118d21347c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.244538] env[61824]: DEBUG nova.network.neutron [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Successfully created port: d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.263058] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98bbd1e-ecd7-4559-b9d1-ebdf89cd2489 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.284960] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] volume-32104921-3bcd-4691-86e7-9c3c0a08ed72/volume-32104921-3bcd-4691-86e7-9c3c0a08ed72.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.287676] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7db80134-a91a-4615-b2f7-f6da3f5506e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.308182] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 803.308182] env[61824]: value = "task-1275491" [ 803.308182] env[61824]: _type = "Task" [ 803.308182] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.318130] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275491, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.358068] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92b49a7-34e8-4c5c-a5e1-1a2502c43ff8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.365937] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30524045-66ce-4d14-8d48-b3973895d50e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.402464] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4caeb378-397f-4bb7-8555-d44fd381751d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.409964] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c55d97-f617-4715-a140-ffc5949cdde2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.414836] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.424470] env[61824]: DEBUG nova.compute.provider_tree [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.679459] env[61824]: INFO nova.compute.manager [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Took 39.06 seconds to build instance. [ 803.737279] env[61824]: DEBUG nova.network.neutron [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.818195] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275491, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.930025] env[61824]: DEBUG nova.scheduler.client.report [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 803.983357] env[61824]: DEBUG nova.network.neutron [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Updating instance_info_cache with network_info: [{"id": "b9d880b8-2784-4570-9726-b6f545278cf2", "address": "fa:16:3e:15:8b:0f", "network": {"id": "1edde1c1-64cc-4185-b09b-44341c22da3e", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1661724750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70e6557f001f486b9c0d7923bd161981", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d880b8-27", "ovs_interfaceid": "b9d880b8-2784-4570-9726-b6f545278cf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.052740] env[61824]: DEBUG nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.079641] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.079907] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.080081] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.080264] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.080410] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.080556] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.080763] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.080918] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.081094] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.081256] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.081424] env[61824]: DEBUG nova.virt.hardware [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.082330] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4dcc7d0-6743-4c6f-9542-4be67317f140 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.090667] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3125fd5d-302f-4e9e-b9cf-9874f7fb010d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.181148] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9f29a559-01e9-4c5a-9e96-3e21af64ff0f tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.866s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.317730] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275491, 'name': ReconfigVM_Task, 'duration_secs': 0.967562} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.317982] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Reconfigured VM instance instance-0000003d to attach disk [datastore2] volume-32104921-3bcd-4691-86e7-9c3c0a08ed72/volume-32104921-3bcd-4691-86e7-9c3c0a08ed72.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.322806] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ebded92-fad6-4350-bfb3-dcff939e4202 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.337737] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 804.337737] env[61824]: value = "task-1275492" [ 804.337737] env[61824]: _type = "Task" [ 804.337737] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.345657] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275492, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.433197] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.433389] env[61824]: DEBUG nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.436547] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.864s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.438026] env[61824]: INFO nova.compute.claims [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 804.486245] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Releasing lock "refresh_cache-0201e783-ed75-4cee-8df4-efac97b4ca19" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.486662] env[61824]: DEBUG nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Instance network_info: |[{"id": "b9d880b8-2784-4570-9726-b6f545278cf2", "address": "fa:16:3e:15:8b:0f", "network": {"id": "1edde1c1-64cc-4185-b09b-44341c22da3e", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1661724750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70e6557f001f486b9c0d7923bd161981", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d880b8-27", "ovs_interfaceid": "b9d880b8-2784-4570-9726-b6f545278cf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 804.487212] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:8b:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ed91b7b-b4ec-486d-ab34-af0afb7ec691', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9d880b8-2784-4570-9726-b6f545278cf2', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 804.498608] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Creating folder: Project (70e6557f001f486b9c0d7923bd161981). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.499279] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-924eaf48-53c2-4d38-824e-ced61cdc3e54 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.512714] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Created folder: Project (70e6557f001f486b9c0d7923bd161981) in parent group-v274074. [ 804.512973] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Creating folder: Instances. Parent ref: group-v274125. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 804.513276] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0be2684e-a5c4-4df6-9c52-cb74aca3c112 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.523018] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Created folder: Instances in parent group-v274125. [ 804.523018] env[61824]: DEBUG oslo.service.loopingcall [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.523117] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 804.523369] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd3287fc-5454-4f19-ad23-5c4038ff6739 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.546873] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 804.546873] env[61824]: value = "task-1275495" [ 804.546873] env[61824]: _type = "Task" [ 804.546873] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.555144] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275495, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.718158] env[61824]: DEBUG nova.compute.manager [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Received event network-changed-b9d880b8-2784-4570-9726-b6f545278cf2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.718158] env[61824]: DEBUG nova.compute.manager [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Refreshing instance network info cache due to event network-changed-b9d880b8-2784-4570-9726-b6f545278cf2. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.718158] env[61824]: DEBUG oslo_concurrency.lockutils [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] Acquiring lock "refresh_cache-0201e783-ed75-4cee-8df4-efac97b4ca19" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.718158] env[61824]: DEBUG oslo_concurrency.lockutils [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] Acquired lock "refresh_cache-0201e783-ed75-4cee-8df4-efac97b4ca19" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.718158] env[61824]: DEBUG nova.network.neutron [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Refreshing network info cache for port b9d880b8-2784-4570-9726-b6f545278cf2 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 804.848482] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275492, 'name': ReconfigVM_Task, 'duration_secs': 0.117293} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.848826] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274093', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'name': 'volume-32104921-3bcd-4691-86e7-9c3c0a08ed72', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6235a9b9-19c0-4917-938a-bca82a2ca05e', 'attached_at': '', 'detached_at': '', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'serial': '32104921-3bcd-4691-86e7-9c3c0a08ed72'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 804.849377] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c97bb6ff-40ca-46e4-8354-308dc23084b3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.856054] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 804.856054] env[61824]: value = "task-1275496" [ 804.856054] env[61824]: _type = "Task" [ 804.856054] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.865154] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275496, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.943085] env[61824]: DEBUG nova.compute.utils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 804.947054] env[61824]: DEBUG nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 804.947312] env[61824]: DEBUG nova.network.neutron [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 805.056484] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275495, 'name': CreateVM_Task, 'duration_secs': 0.417885} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.056685] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.057335] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.057667] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.057983] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.059487] env[61824]: DEBUG nova.policy [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bcb8378feab42ae8666db25003481c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bd0e94b74d74f9898049c9cad364b5b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.061166] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8749fc5-bda0-4801-afd6-46d4e17d92b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.065565] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 805.065565] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52461c5b-839c-4c98-6c4f-8e1278c052fe" [ 805.065565] env[61824]: _type = "Task" [ 805.065565] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.073581] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52461c5b-839c-4c98-6c4f-8e1278c052fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.079031] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "8522fe80-4d6d-4474-ae63-00daa0dc7349" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.079312] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.079439] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "8522fe80-4d6d-4474-ae63-00daa0dc7349-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.079653] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.079796] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.083031] env[61824]: INFO nova.compute.manager [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Terminating instance [ 805.084992] env[61824]: DEBUG nova.compute.manager [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.085193] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 805.085916] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26180d0c-5ace-4f05-8233-7b7f2e45cdaf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.092408] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 805.092624] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-548be201-7bcc-41f5-998a-231f76353b84 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.098245] env[61824]: DEBUG oslo_vmware.api [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 805.098245] env[61824]: value = "task-1275497" [ 805.098245] env[61824]: _type = "Task" [ 805.098245] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.105632] env[61824]: DEBUG oslo_vmware.api [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275497, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.333923] env[61824]: DEBUG nova.network.neutron [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Successfully updated port: d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.367385] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275496, 'name': Rename_Task, 'duration_secs': 0.113585} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.367678] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.367945] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f7b6a6d-fa6f-424a-847d-e1de933ce34f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.375508] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 805.375508] env[61824]: value = "task-1275498" [ 805.375508] env[61824]: _type = "Task" [ 805.375508] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.385965] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.448187] env[61824]: DEBUG nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.451942] env[61824]: DEBUG nova.network.neutron [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Updated VIF entry in instance network info cache for port b9d880b8-2784-4570-9726-b6f545278cf2. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 805.452287] env[61824]: DEBUG nova.network.neutron [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Updating instance_info_cache with network_info: [{"id": "b9d880b8-2784-4570-9726-b6f545278cf2", "address": "fa:16:3e:15:8b:0f", "network": {"id": "1edde1c1-64cc-4185-b09b-44341c22da3e", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1661724750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "70e6557f001f486b9c0d7923bd161981", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9d880b8-27", "ovs_interfaceid": "b9d880b8-2784-4570-9726-b6f545278cf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.541940] env[61824]: DEBUG nova.network.neutron [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Successfully created port: a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.575690] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52461c5b-839c-4c98-6c4f-8e1278c052fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009517} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.576090] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.576990] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 805.576990] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.576990] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.576990] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 805.579592] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21e79953-699c-4ba9-8896-5102c8e62cb3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.588233] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 805.588423] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 805.589164] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c82c5eb2-7a18-4bcc-8054-4e15f2c17a2a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.596749] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 805.596749] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52327306-b190-12c7-f36f-95a2798c39c6" [ 805.596749] env[61824]: _type = "Task" [ 805.596749] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.614115] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52327306-b190-12c7-f36f-95a2798c39c6, 'name': SearchDatastore_Task, 'duration_secs': 0.008042} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.614115] env[61824]: DEBUG oslo_vmware.api [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275497, 'name': PowerOffVM_Task, 'duration_secs': 0.194201} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.614558] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 805.614724] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 805.614932] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47ed0fec-c95c-4487-a27e-0168a9c4cbe0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.617995] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f108e44-f5c3-4645-88af-641dd9c2f3be {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.621897] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 805.621897] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]528586c2-a0f3-d9b8-38b1-152f83ef7821" [ 805.621897] env[61824]: _type = "Task" [ 805.621897] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.633209] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528586c2-a0f3-d9b8-38b1-152f83ef7821, 'name': SearchDatastore_Task, 'duration_secs': 0.008883} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.633517] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.634064] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 0201e783-ed75-4cee-8df4-efac97b4ca19/0201e783-ed75-4cee-8df4-efac97b4ca19.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.634440] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0a5892e-ba61-476f-89a1-ad372d30918e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.640845] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 805.640845] env[61824]: value = "task-1275500" [ 805.640845] env[61824]: _type = "Task" [ 805.640845] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.648554] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275500, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.714325] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d525d71d-c3d3-4eb5-b05f-f986bf744afd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.721443] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f042dd-bcb7-4bee-bbb3-05c7cc35518d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.752765] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214a8a17-b5a6-429e-8acc-f1ab903afd68 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.760033] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923a7004-db4c-47f8-a5f7-783c9e01ef28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.774168] env[61824]: DEBUG nova.compute.provider_tree [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.836652] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.836813] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.836909] env[61824]: DEBUG nova.network.neutron [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.886292] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275498, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.956816] env[61824]: DEBUG oslo_concurrency.lockutils [req-eefa94fa-f4d5-4be5-bfa6-d8baada1b1c0 req-d29dac3f-59a2-4391-866e-ea5c8af4f52d service nova] Releasing lock "refresh_cache-0201e783-ed75-4cee-8df4-efac97b4ca19" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.969530] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 805.969818] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 805.970031] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Deleting the datastore file [datastore2] 8522fe80-4d6d-4474-ae63-00daa0dc7349 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 805.970330] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f3dba2e-bf10-45a4-a95f-fc646289ef11 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.977806] env[61824]: DEBUG oslo_vmware.api [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for the task: (returnval){ [ 805.977806] env[61824]: value = "task-1275501" [ 805.977806] env[61824]: _type = "Task" [ 805.977806] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.988262] env[61824]: DEBUG oslo_vmware.api [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.152616] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275500, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.277351] env[61824]: DEBUG nova.scheduler.client.report [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.372492] env[61824]: DEBUG nova.network.neutron [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.385943] env[61824]: DEBUG oslo_vmware.api [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275498, 'name': PowerOnVM_Task, 'duration_secs': 1.011033} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.386350] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 806.386559] env[61824]: INFO nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Took 5.39 seconds to spawn the instance on the hypervisor. [ 806.386731] env[61824]: DEBUG nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 806.387731] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb351c8-807a-4541-8907-a06a3917bd7f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.464042] env[61824]: DEBUG nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.488754] env[61824]: DEBUG oslo_vmware.api [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275501, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.490907] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.491142] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.491332] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.491554] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.491709] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.491857] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.492068] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.492231] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.492395] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.492576] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.492845] env[61824]: DEBUG nova.virt.hardware [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.493595] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14174516-94ee-4578-9ced-f7567a57d505 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.501150] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436a6e65-d5e5-4037-a5e3-e5a3698ab166 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.506986] env[61824]: DEBUG nova.network.neutron [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Updating instance_info_cache with network_info: [{"id": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "address": "fa:16:3e:ed:92:a2", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e2c604-4e", "ovs_interfaceid": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.652887] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275500, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533699} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.653180] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 0201e783-ed75-4cee-8df4-efac97b4ca19/0201e783-ed75-4cee-8df4-efac97b4ca19.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 806.653389] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.653633] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d762926-d15f-4a39-92a3-285f9090506b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.660623] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 806.660623] env[61824]: value = "task-1275502" [ 806.660623] env[61824]: _type = "Task" [ 806.660623] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.667881] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.753978] env[61824]: DEBUG nova.compute.manager [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Received event network-vif-plugged-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.754223] env[61824]: DEBUG oslo_concurrency.lockutils [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] Acquiring lock "d9bb3172-c058-4e59-a719-d8aa340fba71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.754422] env[61824]: DEBUG oslo_concurrency.lockutils [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.754618] env[61824]: DEBUG oslo_concurrency.lockutils [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.754836] env[61824]: DEBUG nova.compute.manager [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] No waiting events found dispatching network-vif-plugged-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.754997] env[61824]: WARNING nova.compute.manager [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Received unexpected event network-vif-plugged-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce for instance with vm_state building and task_state spawning. [ 806.755176] env[61824]: DEBUG nova.compute.manager [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Received event network-changed-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.755332] env[61824]: DEBUG nova.compute.manager [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Refreshing instance network info cache due to event network-changed-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.755546] env[61824]: DEBUG oslo_concurrency.lockutils [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] Acquiring lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.782900] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.782900] env[61824]: DEBUG nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 806.785702] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.512s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.785702] env[61824]: DEBUG nova.objects.instance [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lazy-loading 'resources' on Instance uuid 23fbfb1e-a716-4a73-a15d-16ff25690e0e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 806.907497] env[61824]: INFO nova.compute.manager [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Took 32.64 seconds to build instance. [ 806.988137] env[61824]: DEBUG oslo_vmware.api [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Task: {'id': task-1275501, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.723655} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.988340] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 806.988570] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 806.988764] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 806.988937] env[61824]: INFO nova.compute.manager [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Took 1.90 seconds to destroy the instance on the hypervisor. [ 806.989188] env[61824]: DEBUG oslo.service.loopingcall [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.989371] env[61824]: DEBUG nova.compute.manager [-] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 806.989465] env[61824]: DEBUG nova.network.neutron [-] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 807.018054] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.018054] env[61824]: DEBUG nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Instance network_info: |[{"id": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "address": "fa:16:3e:ed:92:a2", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e2c604-4e", "ovs_interfaceid": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.018287] env[61824]: DEBUG oslo_concurrency.lockutils [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] Acquired lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.018287] env[61824]: DEBUG nova.network.neutron [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Refreshing network info cache for port d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 807.019102] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:92:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa410d21-2141-45bb-8d0b-16c77304605f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd4e2c604-4e39-49cc-89a3-b4eed4e7d7ce', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.026391] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Creating folder: Project (502b41859ba843838d2188c82f70ae08). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.026812] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0dbf0a8a-198b-437a-88b0-de6b75956636 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.038958] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Created folder: Project (502b41859ba843838d2188c82f70ae08) in parent group-v274074. [ 807.039155] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Creating folder: Instances. Parent ref: group-v274128. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.039380] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a450c38e-db25-44c1-92ca-002107c2fd0a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.049832] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Created folder: Instances in parent group-v274128. [ 807.050062] env[61824]: DEBUG oslo.service.loopingcall [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.050242] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.050432] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b3ff5c1-545e-415a-a555-eefc4e81da46 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.070499] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.070499] env[61824]: value = "task-1275505" [ 807.070499] env[61824]: _type = "Task" [ 807.070499] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.078275] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275505, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.171667] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064838} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.173034] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 807.173153] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3ceaf3-dd6d-40c7-b9f7-f916c4e71926 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.196928] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 0201e783-ed75-4cee-8df4-efac97b4ca19/0201e783-ed75-4cee-8df4-efac97b4ca19.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 807.197500] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf11e78f-3176-4246-a102-ef0b869552d4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.220119] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 807.220119] env[61824]: value = "task-1275506" [ 807.220119] env[61824]: _type = "Task" [ 807.220119] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.227639] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275506, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.261180] env[61824]: DEBUG nova.compute.manager [req-6b0873e6-ced7-4d9f-b438-65472eafb12c req-639ad3ac-d362-4ed9-81d9-bb2668ad4218 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Received event network-vif-deleted-2947ff73-8e81-4e5a-9074-1db5399885ed {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.261406] env[61824]: INFO nova.compute.manager [req-6b0873e6-ced7-4d9f-b438-65472eafb12c req-639ad3ac-d362-4ed9-81d9-bb2668ad4218 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Neutron deleted interface 2947ff73-8e81-4e5a-9074-1db5399885ed; detaching it from the instance and deleting it from the info cache [ 807.261688] env[61824]: DEBUG nova.network.neutron [req-6b0873e6-ced7-4d9f-b438-65472eafb12c req-639ad3ac-d362-4ed9-81d9-bb2668ad4218 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.289333] env[61824]: DEBUG nova.compute.utils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 807.294039] env[61824]: DEBUG nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 807.294298] env[61824]: DEBUG nova.network.neutron [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 807.348753] env[61824]: DEBUG nova.policy [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8e7fc746334a4b7a9153a721b6eafcfb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c09c426f3fd349c190e1058f8ec87634', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 807.412508] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b73470a-88b6-4ae5-987c-168036984ffb tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.910s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.580215] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275505, 'name': CreateVM_Task, 'duration_secs': 0.462946} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.580615] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.584164] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.584164] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.584164] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.584164] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c055bd6-fa4d-418e-be47-6b56531e4065 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.588297] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2bd190-0a3f-4104-b146-8969d7b605b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.593123] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 807.593123] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5203c1d8-ec27-ece9-2b69-9952223e252d" [ 807.593123] env[61824]: _type = "Task" [ 807.593123] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.603434] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8d5b5e-c39a-4b12-9f39-4549889f74dd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.609643] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5203c1d8-ec27-ece9-2b69-9952223e252d, 'name': SearchDatastore_Task, 'duration_secs': 0.011223} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.610304] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.610564] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.610803] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.610950] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.611137] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.611403] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87fc6dd5-3522-4090-b045-009d60f0578c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.637650] env[61824]: DEBUG nova.network.neutron [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Successfully created port: 394196f1-5cde-4f56-bdf3-a4aa3a681a61 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.640387] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a8f1ba-251d-4937-b028-64d9875909b4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.645853] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.645853] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.648180] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9dd1d40-b75a-4299-a13b-58ebd7f95b30 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.654754] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd59c84e-ec89-494b-9615-1797bb6850c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.659114] env[61824]: DEBUG nova.compute.manager [req-b909253c-3f0e-4395-89af-5b9a78d87670 req-35bc0858-f079-4a92-a3b4-f1f9ae292214 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-vif-plugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.659309] env[61824]: DEBUG oslo_concurrency.lockutils [req-b909253c-3f0e-4395-89af-5b9a78d87670 req-35bc0858-f079-4a92-a3b4-f1f9ae292214 service nova] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.659518] env[61824]: DEBUG oslo_concurrency.lockutils [req-b909253c-3f0e-4395-89af-5b9a78d87670 req-35bc0858-f079-4a92-a3b4-f1f9ae292214 service nova] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.659683] env[61824]: DEBUG oslo_concurrency.lockutils [req-b909253c-3f0e-4395-89af-5b9a78d87670 req-35bc0858-f079-4a92-a3b4-f1f9ae292214 service nova] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.659794] env[61824]: DEBUG nova.compute.manager [req-b909253c-3f0e-4395-89af-5b9a78d87670 req-35bc0858-f079-4a92-a3b4-f1f9ae292214 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] No waiting events found dispatching network-vif-plugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 807.659967] env[61824]: WARNING nova.compute.manager [req-b909253c-3f0e-4395-89af-5b9a78d87670 req-35bc0858-f079-4a92-a3b4-f1f9ae292214 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received unexpected event network-vif-plugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b for instance with vm_state building and task_state spawning. [ 807.663311] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 807.663311] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52709c62-0f3e-8bb2-50c8-89c03d09b02f" [ 807.663311] env[61824]: _type = "Task" [ 807.663311] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.672989] env[61824]: DEBUG nova.compute.provider_tree [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.685062] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52709c62-0f3e-8bb2-50c8-89c03d09b02f, 'name': SearchDatastore_Task, 'duration_secs': 0.008441} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.686414] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db324394-7f13-4377-b817-9b8d7151f513 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.691296] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 807.691296] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]522da590-6c7d-b6e6-1632-caf22b7a75ea" [ 807.691296] env[61824]: _type = "Task" [ 807.691296] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.699174] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522da590-6c7d-b6e6-1632-caf22b7a75ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.724360] env[61824]: DEBUG nova.network.neutron [-] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.728556] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275506, 'name': ReconfigVM_Task, 'duration_secs': 0.288522} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.730911] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 0201e783-ed75-4cee-8df4-efac97b4ca19/0201e783-ed75-4cee-8df4-efac97b4ca19.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.731814] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d04b8d3-b88f-4e0b-9fc3-aa23b8961ad3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.740931] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 807.740931] env[61824]: value = "task-1275507" [ 807.740931] env[61824]: _type = "Task" [ 807.740931] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.746502] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275507, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.766028] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7acf4674-f2cd-4928-8422-c67bd050a4fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.773959] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6bf819-58a7-4afd-9131-587fd5f38da6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.801627] env[61824]: DEBUG nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 807.804397] env[61824]: DEBUG nova.compute.manager [req-6b0873e6-ced7-4d9f-b438-65472eafb12c req-639ad3ac-d362-4ed9-81d9-bb2668ad4218 service nova] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Detach interface failed, port_id=2947ff73-8e81-4e5a-9074-1db5399885ed, reason: Instance 8522fe80-4d6d-4474-ae63-00daa0dc7349 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 808.106084] env[61824]: DEBUG nova.network.neutron [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Successfully updated port: a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.180482] env[61824]: DEBUG nova.scheduler.client.report [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.187596] env[61824]: DEBUG nova.network.neutron [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Updated VIF entry in instance network info cache for port d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 808.187596] env[61824]: DEBUG nova.network.neutron [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Updating instance_info_cache with network_info: [{"id": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "address": "fa:16:3e:ed:92:a2", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e2c604-4e", "ovs_interfaceid": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.203202] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522da590-6c7d-b6e6-1632-caf22b7a75ea, 'name': SearchDatastore_Task, 'duration_secs': 0.008081} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.203470] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.203728] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] d9bb3172-c058-4e59-a719-d8aa340fba71/d9bb3172-c058-4e59-a719-d8aa340fba71.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.203976] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d7b07d05-d66a-4910-b589-cf7b10cfc7e3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.211208] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 808.211208] env[61824]: value = "task-1275508" [ 808.211208] env[61824]: _type = "Task" [ 808.211208] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.219487] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.231905] env[61824]: INFO nova.compute.manager [-] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Took 1.24 seconds to deallocate network for instance. [ 808.248889] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275507, 'name': Rename_Task, 'duration_secs': 0.136589} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.249330] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 808.249434] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3fbeae9-a021-4bf2-93da-5794ecaaec75 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.256104] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 808.256104] env[61824]: value = "task-1275509" [ 808.256104] env[61824]: _type = "Task" [ 808.256104] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.266249] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.609242] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.609242] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.609242] env[61824]: DEBUG nova.network.neutron [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 808.685246] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.688558] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.959s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.692543] env[61824]: INFO nova.compute.claims [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 808.693704] env[61824]: DEBUG oslo_concurrency.lockutils [req-b61a5ec2-a5a2-4d7a-ba45-5733ab47976e req-89b95f56-24aa-445a-b8b8-edc2458b5988 service nova] Releasing lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.709496] env[61824]: INFO nova.scheduler.client.report [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Deleted allocations for instance 23fbfb1e-a716-4a73-a15d-16ff25690e0e [ 808.729274] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47951} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.730035] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] d9bb3172-c058-4e59-a719-d8aa340fba71/d9bb3172-c058-4e59-a719-d8aa340fba71.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 808.730365] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.730666] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6acdbd4e-7af5-4b49-b743-843e2080f85b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.738134] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.740466] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 808.740466] env[61824]: value = "task-1275510" [ 808.740466] env[61824]: _type = "Task" [ 808.740466] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.753466] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275510, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.766312] env[61824]: DEBUG oslo_vmware.api [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275509, 'name': PowerOnVM_Task, 'duration_secs': 0.505821} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.766669] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.767077] env[61824]: INFO nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Took 7.12 seconds to spawn the instance on the hypervisor. [ 808.767248] env[61824]: DEBUG nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 808.768063] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af5705a-d40e-4e57-a1b2-70acf8282b01 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.811391] env[61824]: DEBUG nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 808.847841] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 808.847841] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 808.847841] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 808.848248] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 808.848570] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 808.848884] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 808.849578] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 808.850230] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 808.850527] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 808.850825] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 808.853017] env[61824]: DEBUG nova.virt.hardware [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.853017] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa9cff9-50ea-44d4-9de9-f038e6554a1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.863062] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1b6d72-9683-4840-8fe1-9929b167ee71 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.898119] env[61824]: DEBUG nova.compute.manager [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 808.898119] env[61824]: DEBUG nova.compute.manager [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing instance network info cache due to event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 808.898119] env[61824]: DEBUG oslo_concurrency.lockutils [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.149517] env[61824]: DEBUG nova.network.neutron [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 809.190593] env[61824]: DEBUG nova.network.neutron [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Successfully updated port: 394196f1-5cde-4f56-bdf3-a4aa3a681a61 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.225262] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ae80f9c-4954-4242-a690-38018f1d5f09 tempest-ServerMetadataTestJSON-33061249 tempest-ServerMetadataTestJSON-33061249-project-member] Lock "23fbfb1e-a716-4a73-a15d-16ff25690e0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.935s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.253141] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275510, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07686} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.253422] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.254194] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbc8dbf-1d3a-4642-94a0-e97982ea0648 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.277802] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] d9bb3172-c058-4e59-a719-d8aa340fba71/d9bb3172-c058-4e59-a719-d8aa340fba71.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.283271] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7187b5f-aeda-4cac-83c8-112c94520536 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.307021] env[61824]: DEBUG nova.compute.manager [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Received event network-vif-plugged-394196f1-5cde-4f56-bdf3-a4aa3a681a61 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.307021] env[61824]: DEBUG oslo_concurrency.lockutils [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] Acquiring lock "43071be0-eaab-415c-8dcb-948bd9702885-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.307021] env[61824]: DEBUG oslo_concurrency.lockutils [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] Lock "43071be0-eaab-415c-8dcb-948bd9702885-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.307021] env[61824]: DEBUG oslo_concurrency.lockutils [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] Lock "43071be0-eaab-415c-8dcb-948bd9702885-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.307021] env[61824]: DEBUG nova.compute.manager [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] No waiting events found dispatching network-vif-plugged-394196f1-5cde-4f56-bdf3-a4aa3a681a61 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 809.307331] env[61824]: WARNING nova.compute.manager [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Received unexpected event network-vif-plugged-394196f1-5cde-4f56-bdf3-a4aa3a681a61 for instance with vm_state building and task_state spawning. [ 809.307331] env[61824]: DEBUG nova.compute.manager [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Received event network-changed-394196f1-5cde-4f56-bdf3-a4aa3a681a61 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.307331] env[61824]: DEBUG nova.compute.manager [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Refreshing instance network info cache due to event network-changed-394196f1-5cde-4f56-bdf3-a4aa3a681a61. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.307331] env[61824]: DEBUG oslo_concurrency.lockutils [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] Acquiring lock "refresh_cache-43071be0-eaab-415c-8dcb-948bd9702885" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.307331] env[61824]: DEBUG oslo_concurrency.lockutils [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] Acquired lock "refresh_cache-43071be0-eaab-415c-8dcb-948bd9702885" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.307553] env[61824]: DEBUG nova.network.neutron [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Refreshing network info cache for port 394196f1-5cde-4f56-bdf3-a4aa3a681a61 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.308748] env[61824]: INFO nova.compute.manager [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Took 33.49 seconds to build instance. [ 809.315883] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 809.315883] env[61824]: value = "task-1275511" [ 809.315883] env[61824]: _type = "Task" [ 809.315883] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.328875] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275511, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.409422] env[61824]: DEBUG nova.network.neutron [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.693187] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "refresh_cache-43071be0-eaab-415c-8dcb-948bd9702885" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.811666] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fa59b758-bf47-4665-871d-a8121f6b6be4 tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.803s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.831264] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.868668] env[61824]: DEBUG nova.network.neutron [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 809.912104] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.912441] env[61824]: DEBUG nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Instance network_info: |[{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.914859] env[61824]: DEBUG oslo_concurrency.lockutils [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.915075] env[61824]: DEBUG nova.network.neutron [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 809.916297] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:5b:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a66cb8c8-5826-4e02-ad96-b2bbea85d23b', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.929677] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating folder: Project (8bd0e94b74d74f9898049c9cad364b5b). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.932957] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7eb96052-f91d-4982-93be-45e1bea6b350 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.944408] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created folder: Project (8bd0e94b74d74f9898049c9cad364b5b) in parent group-v274074. [ 809.944618] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating folder: Instances. Parent ref: group-v274131. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 809.947231] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76611d03-6392-4ef3-a680-c0cf8fc23588 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.959019] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created folder: Instances in parent group-v274131. [ 809.959019] env[61824]: DEBUG oslo.service.loopingcall [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.959019] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 809.959019] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a7d9b14f-f4f2-4cbf-8bbd-8cef107c6e4b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.987149] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.987149] env[61824]: value = "task-1275514" [ 809.987149] env[61824]: _type = "Task" [ 809.987149] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.997979] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275514, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.105879] env[61824]: DEBUG nova.network.neutron [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.123193] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c357fce-56b9-4502-b9cc-e90209a81eea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.129191] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bd204a-859e-436c-aee0-973d8d500e1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.164748] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f80bcb-a8e9-4f1b-9d57-12592cfa3f87 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.172801] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843e30a7-d464-4003-9b5e-d817c65f4bce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.187197] env[61824]: DEBUG nova.compute.provider_tree [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.207446] env[61824]: DEBUG nova.network.neutron [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updated VIF entry in instance network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 810.207814] env[61824]: DEBUG nova.network.neutron [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.327201] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275511, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.496742] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275514, 'name': CreateVM_Task, 'duration_secs': 0.432692} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.496971] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 810.497983] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.498165] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.498496] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.498706] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf07a24b-20e5-4130-9778-444e1bfccdbc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.503279] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 810.503279] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52019fec-a97b-f4d9-bb17-4190843e26f4" [ 810.503279] env[61824]: _type = "Task" [ 810.503279] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.510900] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52019fec-a97b-f4d9-bb17-4190843e26f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.609071] env[61824]: DEBUG oslo_concurrency.lockutils [req-de52740b-a512-4a22-a400-afd4302fc810 req-a4531285-1175-41cd-b32d-de5e2769f6ef service nova] Releasing lock "refresh_cache-43071be0-eaab-415c-8dcb-948bd9702885" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.609071] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquired lock "refresh_cache-43071be0-eaab-415c-8dcb-948bd9702885" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.609228] env[61824]: DEBUG nova.network.neutron [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 810.690642] env[61824]: DEBUG nova.scheduler.client.report [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.709880] env[61824]: DEBUG oslo_concurrency.lockutils [req-bbb12ecb-68f7-4899-9aa7-460fe5ea0054 req-625e6287-11a2-4b73-8630-39aa76be4074 service nova] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.829395] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275511, 'name': ReconfigVM_Task, 'duration_secs': 1.06967} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.829725] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Reconfigured VM instance instance-0000003f to attach disk [datastore2] d9bb3172-c058-4e59-a719-d8aa340fba71/d9bb3172-c058-4e59-a719-d8aa340fba71.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.830386] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a6a1bdd3-c954-4e87-9f1c-ca0bd06509c7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.840020] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 810.840020] env[61824]: value = "task-1275515" [ 810.840020] env[61824]: _type = "Task" [ 810.840020] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.849686] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275515, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.923465] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "0201e783-ed75-4cee-8df4-efac97b4ca19" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.923750] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.923955] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "0201e783-ed75-4cee-8df4-efac97b4ca19-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.924165] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.924336] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.926762] env[61824]: INFO nova.compute.manager [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Terminating instance [ 810.928750] env[61824]: DEBUG nova.compute.manager [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 810.928942] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.929779] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb4bf1c-32da-4024-89b9-fd3b16d46423 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.937198] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.937622] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65b8edbd-9812-4a80-879d-40f91bcd3e78 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.946718] env[61824]: DEBUG oslo_vmware.api [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 810.946718] env[61824]: value = "task-1275516" [ 810.946718] env[61824]: _type = "Task" [ 810.946718] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.957120] env[61824]: DEBUG oslo_vmware.api [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.977760] env[61824]: DEBUG nova.compute.manager [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Received event network-changed-64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 810.977996] env[61824]: DEBUG nova.compute.manager [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Refreshing instance network info cache due to event network-changed-64000f4d-cb2c-4d63-95b7-97218e98723c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 810.978225] env[61824]: DEBUG oslo_concurrency.lockutils [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] Acquiring lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.978368] env[61824]: DEBUG oslo_concurrency.lockutils [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] Acquired lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.978530] env[61824]: DEBUG nova.network.neutron [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Refreshing network info cache for port 64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.014385] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52019fec-a97b-f4d9-bb17-4190843e26f4, 'name': SearchDatastore_Task, 'duration_secs': 0.009945} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.014648] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.014875] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.015926] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.015926] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.015926] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.015926] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2718cb28-d354-49dd-9bea-dbe8aeafacac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.024101] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.024295] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 811.025288] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61980bca-a5e4-4bc1-87c9-45b1d80248b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.030867] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 811.030867] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5267c0e4-1b52-b705-dff0-4804f7362775" [ 811.030867] env[61824]: _type = "Task" [ 811.030867] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.038775] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5267c0e4-1b52-b705-dff0-4804f7362775, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.154871] env[61824]: DEBUG nova.network.neutron [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 811.195171] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.195694] env[61824]: DEBUG nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 811.199694] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.677s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.199947] env[61824]: DEBUG nova.objects.instance [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lazy-loading 'resources' on Instance uuid cc13d6a5-6b99-40bf-8978-09d284b4ed17 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.348021] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275515, 'name': Rename_Task, 'duration_secs': 0.127919} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.348319] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.348568] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d133a78-764e-4de6-8c3c-0860e8500d11 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.356615] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 811.356615] env[61824]: value = "task-1275517" [ 811.356615] env[61824]: _type = "Task" [ 811.356615] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.363349] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275517, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.380262] env[61824]: DEBUG nova.network.neutron [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Updating instance_info_cache with network_info: [{"id": "394196f1-5cde-4f56-bdf3-a4aa3a681a61", "address": "fa:16:3e:09:9c:95", "network": {"id": "c329c882-6886-4ae9-a5cf-0dddf70c3a8e", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1972603630-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c09c426f3fd349c190e1058f8ec87634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap394196f1-5c", "ovs_interfaceid": "394196f1-5cde-4f56-bdf3-a4aa3a681a61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.456437] env[61824]: DEBUG oslo_vmware.api [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275516, 'name': PowerOffVM_Task, 'duration_secs': 0.192742} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.456679] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 811.456855] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 811.457243] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fcf44576-e67a-4e90-94b0-5d62fb0b1bef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.515315] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 811.515894] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 811.515894] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Deleting the datastore file [datastore2] 0201e783-ed75-4cee-8df4-efac97b4ca19 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 811.515894] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e6e4652-07f7-4a3f-985d-0b12f7987a03 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.523415] env[61824]: DEBUG oslo_vmware.api [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for the task: (returnval){ [ 811.523415] env[61824]: value = "task-1275519" [ 811.523415] env[61824]: _type = "Task" [ 811.523415] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.531563] env[61824]: DEBUG oslo_vmware.api [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275519, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.539255] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5267c0e4-1b52-b705-dff0-4804f7362775, 'name': SearchDatastore_Task, 'duration_secs': 0.008967} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.540124] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d51c149-e527-4ba4-bf95-bba8a2d1db47 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.544821] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 811.544821] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5259769b-2c30-cee3-d702-21e69685ff73" [ 811.544821] env[61824]: _type = "Task" [ 811.544821] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.553787] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5259769b-2c30-cee3-d702-21e69685ff73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.704948] env[61824]: DEBUG nova.compute.utils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 811.706675] env[61824]: DEBUG nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 811.706854] env[61824]: DEBUG nova.network.neutron [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 811.754814] env[61824]: DEBUG nova.policy [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd05e42bfa416428d81c19861b11eea2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '858b5ed95401468fbf97c4000f7302a5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 811.797045] env[61824]: DEBUG nova.network.neutron [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Updated VIF entry in instance network info cache for port 64000f4d-cb2c-4d63-95b7-97218e98723c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 811.797490] env[61824]: DEBUG nova.network.neutron [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Updating instance_info_cache with network_info: [{"id": "64000f4d-cb2c-4d63-95b7-97218e98723c", "address": "fa:16:3e:87:33:70", "network": {"id": "33163902-81f7-4ebc-97ee-1d81884d3375", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-712910259-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.156", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "af46b7fc8900414c91c9f64d41ebe314", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7b4bfde-f109-4f64-adab-e7f06b80685d", "external-id": "nsx-vlan-transportzone-910", "segmentation_id": 910, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64000f4d-cb", "ovs_interfaceid": "64000f4d-cb2c-4d63-95b7-97218e98723c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.867179] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275517, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.882931] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Releasing lock "refresh_cache-43071be0-eaab-415c-8dcb-948bd9702885" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.883379] env[61824]: DEBUG nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Instance network_info: |[{"id": "394196f1-5cde-4f56-bdf3-a4aa3a681a61", "address": "fa:16:3e:09:9c:95", "network": {"id": "c329c882-6886-4ae9-a5cf-0dddf70c3a8e", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1972603630-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c09c426f3fd349c190e1058f8ec87634", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7150f662-0cf1-44f9-ae14-d70f479649b6", "external-id": "nsx-vlan-transportzone-712", "segmentation_id": 712, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap394196f1-5c", "ovs_interfaceid": "394196f1-5cde-4f56-bdf3-a4aa3a681a61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 811.883838] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:9c:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7150f662-0cf1-44f9-ae14-d70f479649b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '394196f1-5cde-4f56-bdf3-a4aa3a681a61', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 811.892895] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Creating folder: Project (c09c426f3fd349c190e1058f8ec87634). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.895565] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ece7bce-450b-4801-a630-f07970ede39a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.905113] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Created folder: Project (c09c426f3fd349c190e1058f8ec87634) in parent group-v274074. [ 811.905310] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Creating folder: Instances. Parent ref: group-v274134. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 811.905572] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95dd5837-5bf0-4c3b-954e-386d541ac7dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.915957] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Created folder: Instances in parent group-v274134. [ 811.916225] env[61824]: DEBUG oslo.service.loopingcall [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.916421] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 811.916627] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e2bfbe9-a800-4cc8-ae2a-6afc157a666b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.937909] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 811.937909] env[61824]: value = "task-1275522" [ 811.937909] env[61824]: _type = "Task" [ 811.937909] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.946041] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275522, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.970392] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0dbdfd-5e6c-4e5c-a670-727d1b45ffff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.977875] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826a7d35-830e-4a53-9275-028df9628b1b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.008149] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275acc24-e2e0-4b1c-ad84-18db2c9c7c5a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.015965] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380ff53b-0c71-4d38-ae57-54c92cc87a69 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.031782] env[61824]: DEBUG nova.compute.provider_tree [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.039980] env[61824]: DEBUG oslo_vmware.api [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Task: {'id': task-1275519, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143258} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.041493] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 812.041493] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 812.041493] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 812.041493] env[61824]: INFO nova.compute.manager [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Took 1.11 seconds to destroy the instance on the hypervisor. [ 812.041769] env[61824]: DEBUG oslo.service.loopingcall [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.041769] env[61824]: DEBUG nova.compute.manager [-] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.041769] env[61824]: DEBUG nova.network.neutron [-] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 812.054511] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5259769b-2c30-cee3-d702-21e69685ff73, 'name': SearchDatastore_Task, 'duration_secs': 0.010209} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.054511] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.054511] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 812.054786] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3a4ee92-c045-405c-835f-2f3082ce9ff5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.061479] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 812.061479] env[61824]: value = "task-1275523" [ 812.061479] env[61824]: _type = "Task" [ 812.061479] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.068765] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275523, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.173812] env[61824]: DEBUG nova.network.neutron [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Successfully created port: ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.214216] env[61824]: DEBUG nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 812.301089] env[61824]: DEBUG oslo_concurrency.lockutils [req-dc2daf4a-2f8c-4181-9704-8cca7774b61b req-edaaa341-2303-4791-977e-5f06a70150f9 service nova] Releasing lock "refresh_cache-6235a9b9-19c0-4917-938a-bca82a2ca05e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.373029] env[61824]: DEBUG nova.compute.manager [req-f6b588d3-5883-4f09-add8-5e4f4852d332 req-48e8386a-ead5-4632-b70f-7b0b98bd1a62 service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Received event network-vif-deleted-b9d880b8-2784-4570-9726-b6f545278cf2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.373029] env[61824]: INFO nova.compute.manager [req-f6b588d3-5883-4f09-add8-5e4f4852d332 req-48e8386a-ead5-4632-b70f-7b0b98bd1a62 service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Neutron deleted interface b9d880b8-2784-4570-9726-b6f545278cf2; detaching it from the instance and deleting it from the info cache [ 812.373626] env[61824]: DEBUG nova.network.neutron [req-f6b588d3-5883-4f09-add8-5e4f4852d332 req-48e8386a-ead5-4632-b70f-7b0b98bd1a62 service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.383190] env[61824]: DEBUG oslo_vmware.api [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275517, 'name': PowerOnVM_Task, 'duration_secs': 0.541501} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.384259] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.384671] env[61824]: INFO nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Took 8.33 seconds to spawn the instance on the hypervisor. [ 812.384993] env[61824]: DEBUG nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.386571] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff488292-d5ea-4dec-8af9-e082c0f60871 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.454148] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275522, 'name': CreateVM_Task, 'duration_secs': 0.325368} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.454148] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 812.454148] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.454148] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.454148] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 812.454148] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a810caff-70a4-4ab3-bdbc-024ab7ee1d97 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.460062] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 812.460062] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52fac4a0-9d59-cc12-1e90-b15f6beccf09" [ 812.460062] env[61824]: _type = "Task" [ 812.460062] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.468892] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fac4a0-9d59-cc12-1e90-b15f6beccf09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.539079] env[61824]: DEBUG nova.scheduler.client.report [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.571288] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275523, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445027} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.571561] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 812.571777] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 812.572030] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad9562eb-9ffe-44b9-98cc-fdf6edb96de0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.578230] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 812.578230] env[61824]: value = "task-1275524" [ 812.578230] env[61824]: _type = "Task" [ 812.578230] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.586285] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275524, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.816035] env[61824]: DEBUG nova.network.neutron [-] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.877777] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdafb35c-4001-4e9e-92c9-333c7ea2ff50 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.888444] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f5402d-63c5-4ff3-9924-107494982fd4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.909195] env[61824]: INFO nova.compute.manager [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Took 30.48 seconds to build instance. [ 812.919241] env[61824]: DEBUG nova.compute.manager [req-f6b588d3-5883-4f09-add8-5e4f4852d332 req-48e8386a-ead5-4632-b70f-7b0b98bd1a62 service nova] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Detach interface failed, port_id=b9d880b8-2784-4570-9726-b6f545278cf2, reason: Instance 0201e783-ed75-4cee-8df4-efac97b4ca19 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 812.970910] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fac4a0-9d59-cc12-1e90-b15f6beccf09, 'name': SearchDatastore_Task, 'duration_secs': 0.034458} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.971268] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.971520] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.971760] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.971922] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.972129] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.972412] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64046348-f87f-46c4-b778-64c31ca86862 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.980879] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.981061] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.981813] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46412516-11a9-4dde-a781-a5e8248fdb18 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.987127] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 812.987127] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f5fcdb-2b06-8ad1-6521-db948f6aee77" [ 812.987127] env[61824]: _type = "Task" [ 812.987127] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.994836] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f5fcdb-2b06-8ad1-6521-db948f6aee77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.041903] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.045388] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.437s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.047567] env[61824]: INFO nova.compute.claims [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.064826] env[61824]: INFO nova.scheduler.client.report [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Deleted allocations for instance cc13d6a5-6b99-40bf-8978-09d284b4ed17 [ 813.088148] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275524, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066603} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.088384] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.089148] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5372b1b1-660f-4445-ad7d-916835989387 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.110974] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.111466] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-591dc791-e5f6-4f2d-bd98-c78b0ee4035f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.130546] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 813.130546] env[61824]: value = "task-1275525" [ 813.130546] env[61824]: _type = "Task" [ 813.130546] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.137914] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275525, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.224853] env[61824]: DEBUG nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 813.251482] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 813.251756] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 813.251931] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.252159] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 813.252341] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.252525] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 813.252771] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 813.252963] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 813.253187] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 813.253384] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 813.253602] env[61824]: DEBUG nova.virt.hardware [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 813.254529] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17df1aac-870c-4959-8fa4-1afcfbb43564 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.263043] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6df824-9d71-4721-862b-06a38b71e07d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.318922] env[61824]: INFO nova.compute.manager [-] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Took 1.28 seconds to deallocate network for instance. [ 813.411525] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dd51086f-4f86-4ec5-bd81-443727762f19 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.964s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.498435] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f5fcdb-2b06-8ad1-6521-db948f6aee77, 'name': SearchDatastore_Task, 'duration_secs': 0.008841} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.499371] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b363ac08-2d8d-42c0-a3c6-200b902bf71c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.504808] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 813.504808] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52ab5f34-8359-4c9e-09e0-58ca2c3184ba" [ 813.504808] env[61824]: _type = "Task" [ 813.504808] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.512757] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ab5f34-8359-4c9e-09e0-58ca2c3184ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.573812] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b2d12df2-7493-462f-83d9-bd9923646f12 tempest-ServerAddressesTestJSON-1407287918 tempest-ServerAddressesTestJSON-1407287918-project-member] Lock "cc13d6a5-6b99-40bf-8978-09d284b4ed17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.531s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.641366] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275525, 'name': ReconfigVM_Task, 'duration_secs': 0.253238} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.641834] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.644643] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0710a8bb-28c4-4fcd-bbfc-4dacef968a85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.649696] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 813.649696] env[61824]: value = "task-1275526" [ 813.649696] env[61824]: _type = "Task" [ 813.649696] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.658980] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275526, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.825792] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.955115] env[61824]: DEBUG nova.compute.manager [req-235b41cd-a0de-4ac6-98e4-4ed4847a79ac req-61d1c5e4-81a7-46de-a5bb-6cd44b7b6509 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Received event network-vif-plugged-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.955115] env[61824]: DEBUG oslo_concurrency.lockutils [req-235b41cd-a0de-4ac6-98e4-4ed4847a79ac req-61d1c5e4-81a7-46de-a5bb-6cd44b7b6509 service nova] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.955115] env[61824]: DEBUG oslo_concurrency.lockutils [req-235b41cd-a0de-4ac6-98e4-4ed4847a79ac req-61d1c5e4-81a7-46de-a5bb-6cd44b7b6509 service nova] Lock "cbded637-56b9-4048-b66e-1edf6311f435-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.955115] env[61824]: DEBUG oslo_concurrency.lockutils [req-235b41cd-a0de-4ac6-98e4-4ed4847a79ac req-61d1c5e4-81a7-46de-a5bb-6cd44b7b6509 service nova] Lock "cbded637-56b9-4048-b66e-1edf6311f435-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.955115] env[61824]: DEBUG nova.compute.manager [req-235b41cd-a0de-4ac6-98e4-4ed4847a79ac req-61d1c5e4-81a7-46de-a5bb-6cd44b7b6509 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] No waiting events found dispatching network-vif-plugged-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 813.955850] env[61824]: WARNING nova.compute.manager [req-235b41cd-a0de-4ac6-98e4-4ed4847a79ac req-61d1c5e4-81a7-46de-a5bb-6cd44b7b6509 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Received unexpected event network-vif-plugged-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa for instance with vm_state building and task_state spawning. [ 814.016384] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ab5f34-8359-4c9e-09e0-58ca2c3184ba, 'name': SearchDatastore_Task, 'duration_secs': 0.009042} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.017025] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.017418] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 43071be0-eaab-415c-8dcb-948bd9702885/43071be0-eaab-415c-8dcb-948bd9702885.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 814.017847] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38f1d5c7-37ce-4e12-a41d-53d5d1f75595 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.024699] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 814.024699] env[61824]: value = "task-1275527" [ 814.024699] env[61824]: _type = "Task" [ 814.024699] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.032984] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275527, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.041916] env[61824]: DEBUG nova.network.neutron [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Successfully updated port: ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.164905] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275526, 'name': Rename_Task, 'duration_secs': 0.135413} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.171160] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 814.171160] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1eea8c1c-e980-4561-bd0a-1fd793af50e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.179913] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 814.179913] env[61824]: value = "task-1275528" [ 814.179913] env[61824]: _type = "Task" [ 814.179913] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.188639] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275528, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.308599] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f1cd53-013d-4fcc-ba4a-c11ad142a302 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.316656] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8fedef-2c3f-4f1e-8cd2-60c5568e1c23 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.353532] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca886057-16c3-4709-bf56-9331ff0113e0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.362855] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc177b5-1b5a-4cea-92e2-43cff7546128 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.378971] env[61824]: DEBUG nova.compute.provider_tree [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.408045] env[61824]: DEBUG nova.compute.manager [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Received event network-changed-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 814.408603] env[61824]: DEBUG nova.compute.manager [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Refreshing instance network info cache due to event network-changed-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 814.408603] env[61824]: DEBUG oslo_concurrency.lockutils [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] Acquiring lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.408728] env[61824]: DEBUG oslo_concurrency.lockutils [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] Acquired lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.408889] env[61824]: DEBUG nova.network.neutron [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Refreshing network info cache for port d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 814.535557] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275527, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453302} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.535902] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 43071be0-eaab-415c-8dcb-948bd9702885/43071be0-eaab-415c-8dcb-948bd9702885.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.536074] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.536333] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a033a59-8899-4e44-8dae-ea95ce387fa0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.543687] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.543846] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquired lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.544310] env[61824]: DEBUG nova.network.neutron [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 814.546074] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 814.546074] env[61824]: value = "task-1275529" [ 814.546074] env[61824]: _type = "Task" [ 814.546074] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.558673] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.694650] env[61824]: DEBUG oslo_vmware.api [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275528, 'name': PowerOnVM_Task, 'duration_secs': 0.496226} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.694919] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.695143] env[61824]: INFO nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Took 8.23 seconds to spawn the instance on the hypervisor. [ 814.695330] env[61824]: DEBUG nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.696158] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d75109-512f-45c2-86c3-f2785935de82 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.882176] env[61824]: DEBUG nova.scheduler.client.report [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.060081] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066219} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.063826] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.065307] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836d7503-066d-445d-898a-b0a62188130d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.093625] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 43071be0-eaab-415c-8dcb-948bd9702885/43071be0-eaab-415c-8dcb-948bd9702885.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.094724] env[61824]: DEBUG nova.network.neutron [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.096642] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de8bdeec-131c-4df4-bd5f-68dfc73d5c86 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.120048] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 815.120048] env[61824]: value = "task-1275530" [ 815.120048] env[61824]: _type = "Task" [ 815.120048] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.128717] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275530, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.226817] env[61824]: INFO nova.compute.manager [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Took 28.28 seconds to build instance. [ 815.266962] env[61824]: DEBUG nova.network.neutron [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updating instance_info_cache with network_info: [{"id": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "address": "fa:16:3e:5e:c4:89", "network": {"id": "dbbd9d7f-c159-48f7-a6b4-543034932a1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1313128773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "858b5ed95401468fbf97c4000f7302a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee9b9520-3d", "ovs_interfaceid": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.389088] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.344s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.389677] env[61824]: DEBUG nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.392673] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.216s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.397471] env[61824]: INFO nova.compute.claims [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.603309] env[61824]: DEBUG nova.network.neutron [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Updated VIF entry in instance network info cache for port d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 815.603309] env[61824]: DEBUG nova.network.neutron [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Updating instance_info_cache with network_info: [{"id": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "address": "fa:16:3e:ed:92:a2", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.208", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd4e2c604-4e", "ovs_interfaceid": "d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.630469] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275530, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.731426] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2df425b1-4c58-49ab-ad1a-c1307481233f tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.242s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.771803] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Releasing lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.771803] env[61824]: DEBUG nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Instance network_info: |[{"id": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "address": "fa:16:3e:5e:c4:89", "network": {"id": "dbbd9d7f-c159-48f7-a6b4-543034932a1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1313128773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "858b5ed95401468fbf97c4000f7302a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee9b9520-3d", "ovs_interfaceid": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 815.772469] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:c4:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee9b9520-3d4f-48fc-bb6d-04a43506c1aa', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.781452] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Creating folder: Project (858b5ed95401468fbf97c4000f7302a5). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.781787] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c963c9d-5976-4ca8-bfe2-f72d999d1f93 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.794078] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Created folder: Project (858b5ed95401468fbf97c4000f7302a5) in parent group-v274074. [ 815.794078] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Creating folder: Instances. Parent ref: group-v274137. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 815.794078] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9249a1d2-f2a8-4420-99ec-9a9a58f3a2b4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.802601] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Created folder: Instances in parent group-v274137. [ 815.802878] env[61824]: DEBUG oslo.service.loopingcall [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.803137] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 815.803356] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3f77e25-2451-4c40-aa63-4d4f45169406 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.823731] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.823731] env[61824]: value = "task-1275533" [ 815.823731] env[61824]: _type = "Task" [ 815.823731] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.831505] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275533, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.898910] env[61824]: DEBUG nova.compute.utils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.906090] env[61824]: DEBUG nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.906090] env[61824]: DEBUG nova.network.neutron [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.006177] env[61824]: DEBUG nova.compute.manager [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Received event network-changed-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.006585] env[61824]: DEBUG nova.compute.manager [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Refreshing instance network info cache due to event network-changed-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 816.007241] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] Acquiring lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.007857] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] Acquired lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.008386] env[61824]: DEBUG nova.network.neutron [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Refreshing network info cache for port ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.033109] env[61824]: DEBUG nova.policy [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1d7f006ab244da48e1bfc38a5859932', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e23c7b193932478096b2fbf49a59fd7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.106036] env[61824]: DEBUG oslo_concurrency.lockutils [req-16f5a57a-91cd-445d-a6ef-71826d8004af req-d19c53ba-8c49-44ad-b030-a0c3a4127a76 service nova] Releasing lock "refresh_cache-d9bb3172-c058-4e59-a719-d8aa340fba71" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.132287] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275530, 'name': ReconfigVM_Task, 'duration_secs': 0.964457} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.132732] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 43071be0-eaab-415c-8dcb-948bd9702885/43071be0-eaab-415c-8dcb-948bd9702885.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.133644] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-59cfd339-2030-44cd-9490-f1cb0230e7c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.143042] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 816.143042] env[61824]: value = "task-1275534" [ 816.143042] env[61824]: _type = "Task" [ 816.143042] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.155087] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275534, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.337672] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275533, 'name': CreateVM_Task, 'duration_secs': 0.301507} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.338047] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.338816] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.339106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.339509] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.339862] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-569b38f0-c495-4f51-a6b6-cbb4d45d7d1a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.345216] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 816.345216] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52398d6d-d6e3-247c-e402-fa379df558d4" [ 816.345216] env[61824]: _type = "Task" [ 816.345216] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.354479] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52398d6d-d6e3-247c-e402-fa379df558d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.406292] env[61824]: DEBUG nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.653291] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275534, 'name': Rename_Task, 'duration_secs': 0.37122} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.653686] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.653996] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b83f607-01ee-4aa3-b925-f787082e217a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.664471] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 816.664471] env[61824]: value = "task-1275535" [ 816.664471] env[61824]: _type = "Task" [ 816.664471] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.674262] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275535, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.690353] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71dc3f1-a419-419b-afa5-0f2ca5603919 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.701921] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb32312-989f-495d-96fc-ce4df815b256 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.706917] env[61824]: DEBUG nova.network.neutron [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Successfully created port: 0660bee3-4b22-4474-ab43-1a3947ada0b1 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.740648] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55a2566-45fd-40e8-8d1c-361d0741019c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.749064] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5f2341-6853-4706-962a-fb9b4df1ddee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.755685] env[61824]: DEBUG nova.network.neutron [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updated VIF entry in instance network info cache for port ee9b9520-3d4f-48fc-bb6d-04a43506c1aa. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 816.756185] env[61824]: DEBUG nova.network.neutron [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updating instance_info_cache with network_info: [{"id": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "address": "fa:16:3e:5e:c4:89", "network": {"id": "dbbd9d7f-c159-48f7-a6b4-543034932a1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1313128773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "858b5ed95401468fbf97c4000f7302a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee9b9520-3d", "ovs_interfaceid": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.767719] env[61824]: DEBUG nova.compute.provider_tree [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.858966] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52398d6d-d6e3-247c-e402-fa379df558d4, 'name': SearchDatastore_Task, 'duration_secs': 0.010313} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.859239] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.859504] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.859774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.859952] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.860187] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.860456] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc3a407a-4ae7-482e-9c90-1ed3446ad021 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.869230] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.869420] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 816.870147] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a5aab7c-298f-4545-a724-70eb74f6f2e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.875135] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 816.875135] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]527404b1-933a-12a0-30eb-42f52d9cb7ba" [ 816.875135] env[61824]: _type = "Task" [ 816.875135] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.882441] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527404b1-933a-12a0-30eb-42f52d9cb7ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.175782] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275535, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.259894] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4ac7a9b-0cd1-4376-a503-8ae0273d7561 req-6ea9c0c0-8d5f-4962-9378-95a33f97b7f0 service nova] Releasing lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.271552] env[61824]: DEBUG nova.scheduler.client.report [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.386302] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527404b1-933a-12a0-30eb-42f52d9cb7ba, 'name': SearchDatastore_Task, 'duration_secs': 0.00848} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.387198] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e7e0bd4-084a-4ef3-ad33-df464a06a466 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.392398] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 817.392398] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f50236-cdeb-ce37-59a9-0ba4a66a0e8f" [ 817.392398] env[61824]: _type = "Task" [ 817.392398] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.400519] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f50236-cdeb-ce37-59a9-0ba4a66a0e8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.419891] env[61824]: DEBUG nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.536042] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.536297] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.536466] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.536620] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.536764] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.536907] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.537122] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.537280] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.537443] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.537631] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.537805] env[61824]: DEBUG nova.virt.hardware [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.538697] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed31301-82bb-4034-b9b9-c64a7ff5b005 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.546332] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5903754b-4c20-4173-828f-83de7123471b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.675331] env[61824]: DEBUG oslo_vmware.api [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275535, 'name': PowerOnVM_Task, 'duration_secs': 0.662825} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.675673] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 817.675863] env[61824]: INFO nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Took 8.86 seconds to spawn the instance on the hypervisor. [ 817.676052] env[61824]: DEBUG nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 817.676834] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14391edb-9875-4821-b685-34e5f88fc35a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.780463] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.781076] env[61824]: DEBUG nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.785347] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.258s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.785347] env[61824]: DEBUG nova.objects.instance [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lazy-loading 'resources' on Instance uuid 3f89fb2d-99c1-4cf7-83aa-db033211a2ce {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 817.906069] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f50236-cdeb-ce37-59a9-0ba4a66a0e8f, 'name': SearchDatastore_Task, 'duration_secs': 0.011005} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.906069] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.906069] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] cbded637-56b9-4048-b66e-1edf6311f435/cbded637-56b9-4048-b66e-1edf6311f435.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 817.906069] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49944426-610b-41ed-ac50-80eeb4f237e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.910785] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 817.910785] env[61824]: value = "task-1275536" [ 817.910785] env[61824]: _type = "Task" [ 817.910785] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.920500] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275536, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.040397] env[61824]: DEBUG nova.compute.manager [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.040397] env[61824]: DEBUG nova.compute.manager [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing instance network info cache due to event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.040397] env[61824]: DEBUG oslo_concurrency.lockutils [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.040397] env[61824]: DEBUG oslo_concurrency.lockutils [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.040397] env[61824]: DEBUG nova.network.neutron [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.200517] env[61824]: INFO nova.compute.manager [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Took 30.65 seconds to build instance. [ 818.288320] env[61824]: DEBUG nova.compute.utils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.290277] env[61824]: DEBUG nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 818.290617] env[61824]: DEBUG nova.network.neutron [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 818.378762] env[61824]: DEBUG nova.policy [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff4fff56a7cc4f19b92786ea07b66e8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bad9de0b032c47429ba879d2d62c6af6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.425324] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275536, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458425} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.425621] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] cbded637-56b9-4048-b66e-1edf6311f435/cbded637-56b9-4048-b66e-1edf6311f435.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.425879] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.426163] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de236b7a-ed2c-44b3-8872-9dd14caedbd4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.437104] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 818.437104] env[61824]: value = "task-1275537" [ 818.437104] env[61824]: _type = "Task" [ 818.437104] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.445905] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275537, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.575178] env[61824]: DEBUG nova.network.neutron [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Successfully updated port: 0660bee3-4b22-4474-ab43-1a3947ada0b1 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.594955] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69597cc6-a728-4ea4-9e0e-492db13adb5d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.604044] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156e59fb-1fd3-4694-a99c-488eb3fad885 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.638584] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911c60cc-3dbe-4417-ad0a-c0ee3f6e0091 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.649917] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56809eb-ce18-443f-9294-a723ce600a44 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.667225] env[61824]: DEBUG nova.compute.provider_tree [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.705423] env[61824]: DEBUG oslo_concurrency.lockutils [None req-97e51574-0bbf-477f-aa29-47a2a6bb0d95 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "43071be0-eaab-415c-8dcb-948bd9702885" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.145s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.799718] env[61824]: DEBUG nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.818997] env[61824]: DEBUG nova.network.neutron [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updated VIF entry in instance network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 818.819380] env[61824]: DEBUG nova.network.neutron [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.821980] env[61824]: DEBUG nova.network.neutron [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Successfully created port: 3734404d-9e67-4190-ab94-d0fd50c732c6 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.947867] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275537, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.232951} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.948394] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 818.949225] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25a56e0-5f45-416e-b7f0-25ca3e76b4cf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.972725] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] cbded637-56b9-4048-b66e-1edf6311f435/cbded637-56b9-4048-b66e-1edf6311f435.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 818.972725] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85ec1075-a4ee-442b-9f14-b92a757b7f75 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.992681] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 818.992681] env[61824]: value = "task-1275538" [ 818.992681] env[61824]: _type = "Task" [ 818.992681] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.000850] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.081498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "refresh_cache-2495fcb4-28cb-49c3-90fe-c84072466287" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.081590] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "refresh_cache-2495fcb4-28cb-49c3-90fe-c84072466287" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.081744] env[61824]: DEBUG nova.network.neutron [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 819.170387] env[61824]: DEBUG nova.scheduler.client.report [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.322553] env[61824]: DEBUG oslo_concurrency.lockutils [req-2d49e38a-c60c-4166-a4c6-b014e4236f3a req-05e892eb-6c86-42a5-a2a7-5edc6e7ec2cc service nova] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.506626] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275538, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.676979] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.891s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.677761] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.453s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.678032] env[61824]: DEBUG nova.objects.instance [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lazy-loading 'resources' on Instance uuid e3e19a04-a29b-4473-833c-d8486703e9ad {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.688754] env[61824]: DEBUG nova.network.neutron [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.708266] env[61824]: INFO nova.scheduler.client.report [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Deleted allocations for instance 3f89fb2d-99c1-4cf7-83aa-db033211a2ce [ 819.810323] env[61824]: DEBUG nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.841111] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.841111] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.841449] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.841724] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.841982] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.842125] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.842407] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.842592] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.842790] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.842974] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.843162] env[61824]: DEBUG nova.virt.hardware [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.844084] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fa8ef8-675a-429b-80b0-4e0e5c66b2e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.856955] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0fa077-13e9-45dd-9450-fba01770311f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.943427] env[61824]: DEBUG nova.network.neutron [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Updating instance_info_cache with network_info: [{"id": "0660bee3-4b22-4474-ab43-1a3947ada0b1", "address": "fa:16:3e:5a:a0:b3", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0660bee3-4b", "ovs_interfaceid": "0660bee3-4b22-4474-ab43-1a3947ada0b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.005100] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275538, 'name': ReconfigVM_Task, 'duration_secs': 0.992369} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.005521] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfigured VM instance instance-00000042 to attach disk [datastore2] cbded637-56b9-4048-b66e-1edf6311f435/cbded637-56b9-4048-b66e-1edf6311f435.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.007369] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ee33a91-74ad-4ae6-af51-12c19100f1ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.012920] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 820.012920] env[61824]: value = "task-1275539" [ 820.012920] env[61824]: _type = "Task" [ 820.012920] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.022177] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275539, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.069700] env[61824]: DEBUG nova.compute.manager [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Received event network-vif-plugged-0660bee3-4b22-4474-ab43-1a3947ada0b1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.070031] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] Acquiring lock "2495fcb4-28cb-49c3-90fe-c84072466287-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.070250] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] Lock "2495fcb4-28cb-49c3-90fe-c84072466287-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.070418] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] Lock "2495fcb4-28cb-49c3-90fe-c84072466287-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.070703] env[61824]: DEBUG nova.compute.manager [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] No waiting events found dispatching network-vif-plugged-0660bee3-4b22-4474-ab43-1a3947ada0b1 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.072245] env[61824]: WARNING nova.compute.manager [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Received unexpected event network-vif-plugged-0660bee3-4b22-4474-ab43-1a3947ada0b1 for instance with vm_state building and task_state spawning. [ 820.072245] env[61824]: DEBUG nova.compute.manager [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Received event network-changed-0660bee3-4b22-4474-ab43-1a3947ada0b1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.072245] env[61824]: DEBUG nova.compute.manager [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Refreshing instance network info cache due to event network-changed-0660bee3-4b22-4474-ab43-1a3947ada0b1. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.072245] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] Acquiring lock "refresh_cache-2495fcb4-28cb-49c3-90fe-c84072466287" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.222355] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c940d6b2-ce59-4c39-aa50-20e434b0da4e tempest-ServerShowV247Test-677901683 tempest-ServerShowV247Test-677901683-project-member] Lock "3f89fb2d-99c1-4cf7-83aa-db033211a2ce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.413s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.310353] env[61824]: DEBUG nova.compute.manager [req-60b8331c-423a-46e2-ab20-fccb1672664a req-1973846c-2c37-499c-908f-ccf131f2f7fb service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Received event network-vif-plugged-3734404d-9e67-4190-ab94-d0fd50c732c6 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.310588] env[61824]: DEBUG oslo_concurrency.lockutils [req-60b8331c-423a-46e2-ab20-fccb1672664a req-1973846c-2c37-499c-908f-ccf131f2f7fb service nova] Acquiring lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.310789] env[61824]: DEBUG oslo_concurrency.lockutils [req-60b8331c-423a-46e2-ab20-fccb1672664a req-1973846c-2c37-499c-908f-ccf131f2f7fb service nova] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.310967] env[61824]: DEBUG oslo_concurrency.lockutils [req-60b8331c-423a-46e2-ab20-fccb1672664a req-1973846c-2c37-499c-908f-ccf131f2f7fb service nova] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.311165] env[61824]: DEBUG nova.compute.manager [req-60b8331c-423a-46e2-ab20-fccb1672664a req-1973846c-2c37-499c-908f-ccf131f2f7fb service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] No waiting events found dispatching network-vif-plugged-3734404d-9e67-4190-ab94-d0fd50c732c6 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.311332] env[61824]: WARNING nova.compute.manager [req-60b8331c-423a-46e2-ab20-fccb1672664a req-1973846c-2c37-499c-908f-ccf131f2f7fb service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Received unexpected event network-vif-plugged-3734404d-9e67-4190-ab94-d0fd50c732c6 for instance with vm_state building and task_state spawning. [ 820.412812] env[61824]: DEBUG nova.network.neutron [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Successfully updated port: 3734404d-9e67-4190-ab94-d0fd50c732c6 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.434664] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36814762-db65-4462-89be-d3da12814ecb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.443693] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c91dc8-5d45-429b-a7c0-ef1c51d48fbe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.447204] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "refresh_cache-2495fcb4-28cb-49c3-90fe-c84072466287" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.447495] env[61824]: DEBUG nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Instance network_info: |[{"id": "0660bee3-4b22-4474-ab43-1a3947ada0b1", "address": "fa:16:3e:5a:a0:b3", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0660bee3-4b", "ovs_interfaceid": "0660bee3-4b22-4474-ab43-1a3947ada0b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.447793] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] Acquired lock "refresh_cache-2495fcb4-28cb-49c3-90fe-c84072466287" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.447985] env[61824]: DEBUG nova.network.neutron [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Refreshing network info cache for port 0660bee3-4b22-4474-ab43-1a3947ada0b1 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.449159] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:a0:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7f41333-42ee-47f3-936c-d6701ab786d2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0660bee3-4b22-4474-ab43-1a3947ada0b1', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.456812] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating folder: Project (e23c7b193932478096b2fbf49a59fd7e). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.460836] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-33e15fb7-5652-48c3-ab40-0010030ab35f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.489460] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc690a0-6a17-42ce-a4ab-0779c5be2a1a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.497981] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef220f52-7e4f-4775-b8b2-953441c98fe0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.502053] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Created folder: Project (e23c7b193932478096b2fbf49a59fd7e) in parent group-v274074. [ 820.502356] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating folder: Instances. Parent ref: group-v274140. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.502852] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f231104b-9aa6-44e1-a203-66128d586653 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.516238] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.516394] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.517782] env[61824]: DEBUG nova.compute.provider_tree [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.526307] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Created folder: Instances in parent group-v274140. [ 820.526535] env[61824]: DEBUG oslo.service.loopingcall [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.526964] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 820.527201] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5120194f-83f2-4503-8044-4675b659e0dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.545705] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275539, 'name': Rename_Task, 'duration_secs': 0.151393} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.546319] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.546757] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e042e7a-7b6d-4436-ba5c-69035255142c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.550614] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.550614] env[61824]: value = "task-1275542" [ 820.550614] env[61824]: _type = "Task" [ 820.550614] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.554561] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 820.554561] env[61824]: value = "task-1275543" [ 820.554561] env[61824]: _type = "Task" [ 820.554561] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.560481] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275542, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.565276] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.719230] env[61824]: DEBUG nova.network.neutron [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Updated VIF entry in instance network info cache for port 0660bee3-4b22-4474-ab43-1a3947ada0b1. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 820.719595] env[61824]: DEBUG nova.network.neutron [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Updating instance_info_cache with network_info: [{"id": "0660bee3-4b22-4474-ab43-1a3947ada0b1", "address": "fa:16:3e:5a:a0:b3", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0660bee3-4b", "ovs_interfaceid": "0660bee3-4b22-4474-ab43-1a3947ada0b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.916066] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "refresh_cache-d6576f45-35f6-48dd-9160-80fc53d8b6b2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.916230] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "refresh_cache-d6576f45-35f6-48dd-9160-80fc53d8b6b2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.916422] env[61824]: DEBUG nova.network.neutron [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.020328] env[61824]: DEBUG nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.026716] env[61824]: DEBUG nova.scheduler.client.report [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.066354] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275542, 'name': CreateVM_Task, 'duration_secs': 0.46632} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.066921] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.068299] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.068299] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.068299] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.068448] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c25b643-83f7-43e2-b524-ec9b55a4b355 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.074027] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275543, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.080013] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 821.080013] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52154df9-64db-fddb-0a04-45c8166f5408" [ 821.080013] env[61824]: _type = "Task" [ 821.080013] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.089982] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52154df9-64db-fddb-0a04-45c8166f5408, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.221853] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d610f15-99fa-403d-84c0-46904fc57531 req-6e42e8bb-1c0c-4ad0-a35e-6f9cce2b3273 service nova] Releasing lock "refresh_cache-2495fcb4-28cb-49c3-90fe-c84072466287" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.468503] env[61824]: DEBUG nova.network.neutron [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.532762] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.536570] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.307s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.536669] env[61824]: DEBUG nova.objects.instance [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lazy-loading 'resources' on Instance uuid f24881d3-a647-42f9-bf82-0065f3425a5f {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.549647] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.559581] env[61824]: INFO nova.scheduler.client.report [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleted allocations for instance e3e19a04-a29b-4473-833c-d8486703e9ad [ 821.565999] env[61824]: DEBUG oslo_vmware.api [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275543, 'name': PowerOnVM_Task, 'duration_secs': 0.802814} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.566336] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.566535] env[61824]: INFO nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Took 8.34 seconds to spawn the instance on the hypervisor. [ 821.566690] env[61824]: DEBUG nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.567538] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d729b4-803c-4c4d-a557-dbb2135e8dc0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.590431] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52154df9-64db-fddb-0a04-45c8166f5408, 'name': SearchDatastore_Task, 'duration_secs': 0.009677} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.590431] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.590431] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.590431] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.590638] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.590777] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.590929] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca41629c-9425-4256-a261-a4ec20417298 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.599108] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.599322] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 821.600071] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc62f6bf-48a7-4c44-baa9-267b0e6d0233 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.605935] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 821.605935] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52391e63-7680-5e30-9a4a-9f224a281d39" [ 821.605935] env[61824]: _type = "Task" [ 821.605935] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.615932] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52391e63-7680-5e30-9a4a-9f224a281d39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.654054] env[61824]: DEBUG nova.network.neutron [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Updating instance_info_cache with network_info: [{"id": "3734404d-9e67-4190-ab94-d0fd50c732c6", "address": "fa:16:3e:1d:da:60", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3734404d-9e", "ovs_interfaceid": "3734404d-9e67-4190-ab94-d0fd50c732c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.673591] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "43071be0-eaab-415c-8dcb-948bd9702885" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.673848] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "43071be0-eaab-415c-8dcb-948bd9702885" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.674259] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "43071be0-eaab-415c-8dcb-948bd9702885-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.674259] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "43071be0-eaab-415c-8dcb-948bd9702885-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.674397] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "43071be0-eaab-415c-8dcb-948bd9702885-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.676421] env[61824]: INFO nova.compute.manager [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Terminating instance [ 821.678422] env[61824]: DEBUG nova.compute.manager [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 821.678664] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 821.679476] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d751a576-d83c-4647-b8ac-a253b7f362b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.687964] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 821.688237] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcd319fa-61ea-4863-bce2-f14ba85d0008 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.694507] env[61824]: DEBUG oslo_vmware.api [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 821.694507] env[61824]: value = "task-1275544" [ 821.694507] env[61824]: _type = "Task" [ 821.694507] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.702801] env[61824]: DEBUG oslo_vmware.api [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.070385] env[61824]: DEBUG oslo_concurrency.lockutils [None req-05fc32a1-8fe1-4cfb-ae9a-190380e58d6a tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "e3e19a04-a29b-4473-833c-d8486703e9ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.057s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.090510] env[61824]: INFO nova.compute.manager [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Took 29.38 seconds to build instance. [ 822.130133] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52391e63-7680-5e30-9a4a-9f224a281d39, 'name': SearchDatastore_Task, 'duration_secs': 0.008882} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.130617] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0179477a-e7ab-4166-9e7b-801d91eec8dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.139502] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 822.139502] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c82275-b452-e8cf-d1bc-77da60c56486" [ 822.139502] env[61824]: _type = "Task" [ 822.139502] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.154890] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c82275-b452-e8cf-d1bc-77da60c56486, 'name': SearchDatastore_Task, 'duration_secs': 0.010783} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.157819] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.158722] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 2495fcb4-28cb-49c3-90fe-c84072466287/2495fcb4-28cb-49c3-90fe-c84072466287.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 822.159272] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "refresh_cache-d6576f45-35f6-48dd-9160-80fc53d8b6b2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.159953] env[61824]: DEBUG nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Instance network_info: |[{"id": "3734404d-9e67-4190-ab94-d0fd50c732c6", "address": "fa:16:3e:1d:da:60", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3734404d-9e", "ovs_interfaceid": "3734404d-9e67-4190-ab94-d0fd50c732c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.160298] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c2517bf8-753b-4384-bee2-14cf91873dde {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.162710] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:da:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6076d24d-3c8e-4bbb-ba96-a08fb27a73cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3734404d-9e67-4190-ab94-d0fd50c732c6', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.172021] env[61824]: DEBUG oslo.service.loopingcall [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.172104] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.173109] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-17d14f11-fa0c-4704-b3bf-518945e7c29a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.193904] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 822.193904] env[61824]: value = "task-1275545" [ 822.193904] env[61824]: _type = "Task" [ 822.193904] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.202951] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.202951] env[61824]: value = "task-1275546" [ 822.202951] env[61824]: _type = "Task" [ 822.202951] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.213505] env[61824]: DEBUG oslo_vmware.api [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.213789] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275545, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.222203] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275546, 'name': CreateVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.338606] env[61824]: DEBUG nova.compute.manager [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Received event network-changed-3734404d-9e67-4190-ab94-d0fd50c732c6 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.339011] env[61824]: DEBUG nova.compute.manager [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Refreshing instance network info cache due to event network-changed-3734404d-9e67-4190-ab94-d0fd50c732c6. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.339452] env[61824]: DEBUG oslo_concurrency.lockutils [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] Acquiring lock "refresh_cache-d6576f45-35f6-48dd-9160-80fc53d8b6b2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.339669] env[61824]: DEBUG oslo_concurrency.lockutils [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] Acquired lock "refresh_cache-d6576f45-35f6-48dd-9160-80fc53d8b6b2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.339811] env[61824]: DEBUG nova.network.neutron [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Refreshing network info cache for port 3734404d-9e67-4190-ab94-d0fd50c732c6 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.360103] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee74184a-402e-4e96-aca8-b05e6070aed9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.368780] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e63e9ae-74ee-4f59-b3ee-010163a8b35a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.407319] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617b44d6-6a8d-4bc0-bccc-fc612251724e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.416691] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54c31f0c-03cf-4236-b632-4eb58017445b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.431661] env[61824]: DEBUG nova.compute.provider_tree [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.592620] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6df1b2dd-00d9-4973-ba88-9d9ebfbbe72e tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.855s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.713637] env[61824]: DEBUG oslo_vmware.api [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275544, 'name': PowerOffVM_Task, 'duration_secs': 0.912204} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.713983] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275545, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489645} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.714500] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.714675] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 822.714919] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 2495fcb4-28cb-49c3-90fe-c84072466287/2495fcb4-28cb-49c3-90fe-c84072466287.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 822.715202] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.715342] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f279ee16-7471-4d48-a7a6-bb7b50d7a547 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.716703] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ed7ac57-7652-4317-845a-0691d6c4609c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.721226] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275546, 'name': CreateVM_Task, 'duration_secs': 0.402229} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.721940] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.722636] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.722793] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.723106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.723332] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-045114f6-7475-4405-9af8-72370525c494 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.725806] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 822.725806] env[61824]: value = "task-1275548" [ 822.725806] env[61824]: _type = "Task" [ 822.725806] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.729851] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 822.729851] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52713f49-3537-2a10-3283-c634aa966add" [ 822.729851] env[61824]: _type = "Task" [ 822.729851] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.736103] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275548, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.740847] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52713f49-3537-2a10-3283-c634aa966add, 'name': SearchDatastore_Task, 'duration_secs': 0.008579} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.741110] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.741328] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.741548] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.741707] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.741875] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.742120] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e85362f-f9b9-4443-bcce-ed79d3a503d4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.749026] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.749212] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.749916] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4af00bbd-789b-41f6-a13c-0ac7fb57b347 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.754428] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 822.754428] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5273f6e7-af4a-f95a-a3ec-78337413c809" [ 822.754428] env[61824]: _type = "Task" [ 822.754428] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.761706] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5273f6e7-af4a-f95a-a3ec-78337413c809, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.832542] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 822.832856] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 822.833069] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Deleting the datastore file [datastore1] 43071be0-eaab-415c-8dcb-948bd9702885 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.833326] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81296207-4f1d-47e7-90fb-bc46bda0d1b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.839393] env[61824]: DEBUG oslo_vmware.api [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for the task: (returnval){ [ 822.839393] env[61824]: value = "task-1275549" [ 822.839393] env[61824]: _type = "Task" [ 822.839393] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.849273] env[61824]: DEBUG oslo_vmware.api [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275549, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.935119] env[61824]: DEBUG nova.scheduler.client.report [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.059173] env[61824]: DEBUG nova.network.neutron [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Updated VIF entry in instance network info cache for port 3734404d-9e67-4190-ab94-d0fd50c732c6. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.059521] env[61824]: DEBUG nova.network.neutron [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Updating instance_info_cache with network_info: [{"id": "3734404d-9e67-4190-ab94-d0fd50c732c6", "address": "fa:16:3e:1d:da:60", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3734404d-9e", "ovs_interfaceid": "3734404d-9e67-4190-ab94-d0fd50c732c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.235688] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275548, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.263484] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5273f6e7-af4a-f95a-a3ec-78337413c809, 'name': SearchDatastore_Task, 'duration_secs': 0.008419} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.264264] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db628ede-f0cc-4e40-b380-277cdf7c1074 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.269264] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 823.269264] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5213934a-1069-9d56-5270-55c09b8d6159" [ 823.269264] env[61824]: _type = "Task" [ 823.269264] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.277018] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5213934a-1069-9d56-5270-55c09b8d6159, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.349479] env[61824]: DEBUG oslo_vmware.api [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Task: {'id': task-1275549, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134297} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.349722] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.349922] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.350087] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.350260] env[61824]: INFO nova.compute.manager [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Took 1.67 seconds to destroy the instance on the hypervisor. [ 823.350493] env[61824]: DEBUG oslo.service.loopingcall [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.350730] env[61824]: DEBUG nova.compute.manager [-] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.350839] env[61824]: DEBUG nova.network.neutron [-] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 823.442904] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.445150] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.835s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.446881] env[61824]: INFO nova.compute.claims [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.479823] env[61824]: INFO nova.scheduler.client.report [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Deleted allocations for instance f24881d3-a647-42f9-bf82-0065f3425a5f [ 823.562427] env[61824]: DEBUG oslo_concurrency.lockutils [req-0faffb38-f7fa-4625-8c46-3415f99c23b1 req-a9998c3d-6a9f-4dae-97f5-15650e6ae671 service nova] Releasing lock "refresh_cache-d6576f45-35f6-48dd-9160-80fc53d8b6b2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.651897] env[61824]: DEBUG nova.compute.manager [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Received event network-changed-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.652115] env[61824]: DEBUG nova.compute.manager [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Refreshing instance network info cache due to event network-changed-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 823.652328] env[61824]: DEBUG oslo_concurrency.lockutils [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] Acquiring lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.652469] env[61824]: DEBUG oslo_concurrency.lockutils [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] Acquired lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.652628] env[61824]: DEBUG nova.network.neutron [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Refreshing network info cache for port ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.739273] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275548, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.781660] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5213934a-1069-9d56-5270-55c09b8d6159, 'name': SearchDatastore_Task, 'duration_secs': 0.009392} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.781945] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.782212] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] d6576f45-35f6-48dd-9160-80fc53d8b6b2/d6576f45-35f6-48dd-9160-80fc53d8b6b2.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.782668] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fcc18cef-d772-42c4-817e-edaf74a04dce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.790660] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 823.790660] env[61824]: value = "task-1275550" [ 823.790660] env[61824]: _type = "Task" [ 823.790660] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.803134] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275550, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.988604] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f5dba7e9-daa2-44a1-b738-3bd2f67bdf6c tempest-MultipleCreateTestJSON-1288123805 tempest-MultipleCreateTestJSON-1288123805-project-member] Lock "f24881d3-a647-42f9-bf82-0065f3425a5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.040s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.100554] env[61824]: DEBUG nova.network.neutron [-] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.239386] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275548, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.071157} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.243139] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.243139] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c64833-beec-48e3-b241-18b8f0234ee9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.268412] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 2495fcb4-28cb-49c3-90fe-c84072466287/2495fcb4-28cb-49c3-90fe-c84072466287.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.269674] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97c18437-6a72-434e-870f-516738e3e380 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.298105] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 824.298105] env[61824]: value = "task-1275551" [ 824.298105] env[61824]: _type = "Task" [ 824.298105] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.303922] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275550, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452973} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.303922] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] d6576f45-35f6-48dd-9160-80fc53d8b6b2/d6576f45-35f6-48dd-9160-80fc53d8b6b2.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.304285] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.304450] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-228c2321-8b43-4944-a84e-fe55857e6191 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.311933] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275551, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.313374] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 824.313374] env[61824]: value = "task-1275552" [ 824.313374] env[61824]: _type = "Task" [ 824.313374] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.435074] env[61824]: DEBUG nova.compute.manager [req-03535df2-57db-495e-85cd-40e12abd4296 req-7ca65b77-74cc-4cfb-93ef-879db91b640b service nova] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Received event network-vif-deleted-394196f1-5cde-4f56-bdf3-a4aa3a681a61 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.543794] env[61824]: DEBUG nova.network.neutron [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updated VIF entry in instance network info cache for port ee9b9520-3d4f-48fc-bb6d-04a43506c1aa. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.544545] env[61824]: DEBUG nova.network.neutron [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updating instance_info_cache with network_info: [{"id": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "address": "fa:16:3e:5e:c4:89", "network": {"id": "dbbd9d7f-c159-48f7-a6b4-543034932a1f", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1313128773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "858b5ed95401468fbf97c4000f7302a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee9b9520-3d", "ovs_interfaceid": "ee9b9520-3d4f-48fc-bb6d-04a43506c1aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.604058] env[61824]: INFO nova.compute.manager [-] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Took 1.25 seconds to deallocate network for instance. [ 824.680801] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3e3394-0885-49c7-ba2b-9425f3d4616e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.691228] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19309cc3-66a1-4467-b01b-421b6a9f7576 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.728751] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f405652-abc2-4204-bdc9-2e6936d4556d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.735113] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffaae94f-735b-412b-9b55-5a560e1945f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.750601] env[61824]: DEBUG nova.compute.provider_tree [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.815638] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275551, 'name': ReconfigVM_Task, 'duration_secs': 0.295974} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.821841] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 2495fcb4-28cb-49c3-90fe-c84072466287/2495fcb4-28cb-49c3-90fe-c84072466287.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.821841] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e578a615-76b2-4671-a0df-8f90d17619e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.834509] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275552, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069662} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.834509] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.834509] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 824.834509] env[61824]: value = "task-1275553" [ 824.834509] env[61824]: _type = "Task" [ 824.834509] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.835117] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff3b85d-713e-46d5-a04a-9dcc0571fa7e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.846486] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275553, 'name': Rename_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.865752] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] d6576f45-35f6-48dd-9160-80fc53d8b6b2/d6576f45-35f6-48dd-9160-80fc53d8b6b2.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.866113] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24760771-f1b6-40dc-a650-c5f2b99b9831 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.888905] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 824.888905] env[61824]: value = "task-1275554" [ 824.888905] env[61824]: _type = "Task" [ 824.888905] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.897624] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275554, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.047855] env[61824]: DEBUG oslo_concurrency.lockutils [req-bdd6eb2a-da3b-478c-9c74-651ff25af192 req-da4f3d91-d87b-4a52-97af-eac927fb7d3e service nova] Releasing lock "refresh_cache-cbded637-56b9-4048-b66e-1edf6311f435" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.114314] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.253757] env[61824]: DEBUG nova.scheduler.client.report [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.350406] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275553, 'name': Rename_Task, 'duration_secs': 0.135883} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.351069] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.351069] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3537b3f9-8e1f-461f-abe7-b4555e5fdde7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.356716] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 825.356716] env[61824]: value = "task-1275555" [ 825.356716] env[61824]: _type = "Task" [ 825.356716] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.364799] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.399787] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.759160] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.314s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.759688] env[61824]: DEBUG nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 825.763079] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 22.348s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.868259] env[61824]: DEBUG oslo_vmware.api [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275555, 'name': PowerOnVM_Task, 'duration_secs': 0.471814} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.868782] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.869043] env[61824]: INFO nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Took 8.45 seconds to spawn the instance on the hypervisor. [ 825.869317] env[61824]: DEBUG nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.871366] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f67b712-9e6c-49de-a61d-14b07b175f91 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.898813] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.270910] env[61824]: DEBUG nova.compute.utils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.285022] env[61824]: INFO nova.compute.claims [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.288641] env[61824]: DEBUG nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.289360] env[61824]: DEBUG nova.network.neutron [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.344031] env[61824]: DEBUG nova.policy [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '316c089785a44aa897040f3c0ef7a071', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0924bfa41b5e42d68818557d2b0ace7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.362046] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.362332] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.403027] env[61824]: INFO nova.compute.manager [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Took 28.81 seconds to build instance. [ 826.410435] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275554, 'name': ReconfigVM_Task, 'duration_secs': 1.096132} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.410732] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Reconfigured VM instance instance-00000044 to attach disk [datastore1] d6576f45-35f6-48dd-9160-80fc53d8b6b2/d6576f45-35f6-48dd-9160-80fc53d8b6b2.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.411363] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-90435fc4-e544-48b3-be36-966c379b490a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.418515] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 826.418515] env[61824]: value = "task-1275556" [ 826.418515] env[61824]: _type = "Task" [ 826.418515] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.428077] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275556, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.794727] env[61824]: DEBUG nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 826.806547] env[61824]: INFO nova.compute.resource_tracker [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating resource usage from migration de82b366-9a8e-489e-bd6f-2729751101dc [ 826.869524] env[61824]: DEBUG nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.886902] env[61824]: DEBUG nova.network.neutron [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Successfully created port: 9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.906086] env[61824]: DEBUG oslo_concurrency.lockutils [None req-24483d2d-c737-48a1-8560-bd4dfb33f123 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "2495fcb4-28cb-49c3-90fe-c84072466287" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.451s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.938024] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275556, 'name': Rename_Task, 'duration_secs': 0.352595} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.938024] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.938024] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-538c155c-194f-4e82-8a05-7800df0c3ada {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.943370] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 826.943370] env[61824]: value = "task-1275557" [ 826.943370] env[61824]: _type = "Task" [ 826.943370] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.959650] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275557, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.106986] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd9711c-8327-48c7-b7d4-ac5c8e7e8632 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.115303] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a470ac6-0cf2-48b2-a97b-85c9a7fc9a5d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.153096] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792edb85-a802-4f56-a4c8-685ad8b7e08d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.160415] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34297a5d-a5ca-4046-a16f-3d98becfcc42 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.176253] env[61824]: DEBUG nova.compute.provider_tree [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.395361] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.455540] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275557, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.681246] env[61824]: DEBUG nova.scheduler.client.report [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.822110] env[61824]: DEBUG nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 827.849803] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.850165] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.850331] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.850514] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.850690] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.850848] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.851063] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.851272] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.851472] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.851658] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.851816] env[61824]: DEBUG nova.virt.hardware [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.852982] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aff4732-2cf4-4b08-bf1b-a43b3f0806b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.861071] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7047db7-e81c-42a7-8340-413123df1385 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.955180] env[61824]: DEBUG oslo_vmware.api [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275557, 'name': PowerOnVM_Task, 'duration_secs': 0.983057} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.955339] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.956157] env[61824]: INFO nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Took 8.15 seconds to spawn the instance on the hypervisor. [ 827.956157] env[61824]: DEBUG nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 827.956586] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89301b4a-5df2-4de3-b009-4623f934df37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.187074] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.424s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.187311] env[61824]: INFO nova.compute.manager [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Migrating [ 828.187544] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.187732] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.190463] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.451s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.190463] env[61824]: DEBUG nova.objects.instance [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lazy-loading 'resources' on Instance uuid 8522fe80-4d6d-4474-ae63-00daa0dc7349 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 828.401923] env[61824]: DEBUG nova.compute.manager [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.403401] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780fe241-2af7-4818-aad0-183834968bce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.411016] env[61824]: DEBUG nova.compute.manager [req-d7852fa4-8a3c-4b6d-9eda-09bf7f4e8849 req-bc700580-924d-4f10-b18e-b5736e07b5f1 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-vif-plugged-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.411432] env[61824]: DEBUG oslo_concurrency.lockutils [req-d7852fa4-8a3c-4b6d-9eda-09bf7f4e8849 req-bc700580-924d-4f10-b18e-b5736e07b5f1 service nova] Acquiring lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.411802] env[61824]: DEBUG oslo_concurrency.lockutils [req-d7852fa4-8a3c-4b6d-9eda-09bf7f4e8849 req-bc700580-924d-4f10-b18e-b5736e07b5f1 service nova] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.412139] env[61824]: DEBUG oslo_concurrency.lockutils [req-d7852fa4-8a3c-4b6d-9eda-09bf7f4e8849 req-bc700580-924d-4f10-b18e-b5736e07b5f1 service nova] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.413398] env[61824]: DEBUG nova.compute.manager [req-d7852fa4-8a3c-4b6d-9eda-09bf7f4e8849 req-bc700580-924d-4f10-b18e-b5736e07b5f1 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] No waiting events found dispatching network-vif-plugged-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 828.413398] env[61824]: WARNING nova.compute.manager [req-d7852fa4-8a3c-4b6d-9eda-09bf7f4e8849 req-bc700580-924d-4f10-b18e-b5736e07b5f1 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received unexpected event network-vif-plugged-9baa0723-69a3-42a4-86c5-38ca3d336970 for instance with vm_state building and task_state spawning. [ 828.476381] env[61824]: INFO nova.compute.manager [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Took 30.32 seconds to build instance. [ 828.513522] env[61824]: DEBUG nova.network.neutron [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Successfully updated port: 9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.692163] env[61824]: INFO nova.compute.rpcapi [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 828.692654] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.924160] env[61824]: INFO nova.compute.manager [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] instance snapshotting [ 828.926174] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ffe089-18be-4481-a90b-769503e097b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.950749] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817cde20-5846-4f7d-954e-f9fea31b39a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.964330] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8798bec-b6dc-4d11-8b59-f989348bd29f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.972855] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b3b752-b646-4cc2-bf91-c907d397cffc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.979484] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8fdb8595-3315-416f-a8f6-ecbc1f9fc83d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.170s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.008830] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9838d356-a1e7-4b31-a685-8364d0d8075b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.016489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.016685] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.016796] env[61824]: DEBUG nova.network.neutron [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.018999] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e5a4a9-c439-4685-923b-09bb99156e30 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.035639] env[61824]: DEBUG nova.compute.provider_tree [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.212808] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.212808] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.212808] env[61824]: DEBUG nova.network.neutron [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.402097] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.402381] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.464182] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Creating Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 829.464548] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c3d8e048-fb6e-4411-a86d-659328f1fceb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.473136] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 829.473136] env[61824]: value = "task-1275558" [ 829.473136] env[61824]: _type = "Task" [ 829.473136] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.482214] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275558, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.538088] env[61824]: DEBUG nova.scheduler.client.report [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.562698] env[61824]: DEBUG nova.network.neutron [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.778498] env[61824]: DEBUG nova.network.neutron [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.905201] env[61824]: DEBUG nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 829.985723] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275558, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.011926] env[61824]: DEBUG nova.network.neutron [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance_info_cache with network_info: [{"id": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "address": "fa:16:3e:ce:e9:8d", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape95afb5b-11", "ovs_interfaceid": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.043837] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.048266] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.222s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.048462] env[61824]: DEBUG nova.objects.instance [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lazy-loading 'resources' on Instance uuid 0201e783-ed75-4cee-8df4-efac97b4ca19 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.078930] env[61824]: INFO nova.scheduler.client.report [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Deleted allocations for instance 8522fe80-4d6d-4474-ae63-00daa0dc7349 [ 830.282635] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.282635] env[61824]: DEBUG nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Instance network_info: |[{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 830.282875] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:98:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6edb8eae-1113-49d0-84f7-9fd9f82b26fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9baa0723-69a3-42a4-86c5-38ca3d336970', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.291556] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Creating folder: Project (0924bfa41b5e42d68818557d2b0ace7d). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.291951] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-463ef34d-e26a-4ded-a048-6e2dac896802 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.305584] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Created folder: Project (0924bfa41b5e42d68818557d2b0ace7d) in parent group-v274074. [ 830.305781] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Creating folder: Instances. Parent ref: group-v274144. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.306067] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbe028f0-505b-43b3-a3e2-c27b8f785382 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.315115] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Created folder: Instances in parent group-v274144. [ 830.315517] env[61824]: DEBUG oslo.service.loopingcall [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.315831] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.316184] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3aba1c35-f673-4452-aa76-f5e1c27d3e20 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.351977] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.351977] env[61824]: value = "task-1275561" [ 830.351977] env[61824]: _type = "Task" [ 830.351977] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.365301] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275561, 'name': CreateVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.432150] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.435822] env[61824]: DEBUG nova.compute.manager [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.436407] env[61824]: DEBUG nova.compute.manager [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing instance network info cache due to event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.436740] env[61824]: DEBUG oslo_concurrency.lockutils [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.436840] env[61824]: DEBUG oslo_concurrency.lockutils [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.437060] env[61824]: DEBUG nova.network.neutron [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.486840] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275558, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.518880] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.590468] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a098ba6e-6e9b-414e-94ef-cf969d7c1cd9 tempest-ServerGroupTestJSON-1011196551 tempest-ServerGroupTestJSON-1011196551-project-member] Lock "8522fe80-4d6d-4474-ae63-00daa0dc7349" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.511s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.800894] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7967dd38-572c-42c7-b3a8-62157d4e9556 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.808806] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4acbf85b-f0b3-4b21-9043-9ce35571f762 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.842255] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bd6568-ef60-494d-8504-e8ccae6097a5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.850768] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52d82c8-d6cf-4e2b-b1d1-1c74f4d07418 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.868108] env[61824]: DEBUG nova.compute.provider_tree [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.873227] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275561, 'name': CreateVM_Task, 'duration_secs': 0.486382} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.873636] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.874335] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.874496] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.875105] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.875367] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33081517-1e26-4c4f-9165-9bfd252621d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.882257] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 830.882257] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]520aef85-526a-66ed-1f7f-4b801cc64b6b" [ 830.882257] env[61824]: _type = "Task" [ 830.882257] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.892501] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520aef85-526a-66ed-1f7f-4b801cc64b6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.986528] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275558, 'name': CreateSnapshot_Task, 'duration_secs': 1.152428} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.987136] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Created Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 830.988017] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d103cf4a-cebe-435b-9ca6-af26f63c85c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.376302] env[61824]: DEBUG nova.scheduler.client.report [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.399089] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520aef85-526a-66ed-1f7f-4b801cc64b6b, 'name': SearchDatastore_Task, 'duration_secs': 0.011212} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.399230] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.399415] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.399641] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.400089] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.400089] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.400233] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-238a19ff-7115-4aed-9d99-70a1271840a9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.411130] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.411329] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.412069] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a3b5d21-81ec-4597-8c7c-0ac2d4607bfa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.420480] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 831.420480] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52993a6a-87aa-7702-3ead-964518d66aea" [ 831.420480] env[61824]: _type = "Task" [ 831.420480] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.434777] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52993a6a-87aa-7702-3ead-964518d66aea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.509706] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Creating linked-clone VM from snapshot {{(pid=61824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 831.510620] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-64c384a9-d480-40f9-bf20-2b9f3c7c5c85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.521277] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 831.521277] env[61824]: value = "task-1275562" [ 831.521277] env[61824]: _type = "Task" [ 831.521277] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.536943] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275562, 'name': CloneVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.641915] env[61824]: DEBUG nova.network.neutron [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updated VIF entry in instance network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.642446] env[61824]: DEBUG nova.network.neutron [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.891069] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.896031] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.344s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.896031] env[61824]: INFO nova.compute.claims [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.920714] env[61824]: INFO nova.scheduler.client.report [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Deleted allocations for instance 0201e783-ed75-4cee-8df4-efac97b4ca19 [ 831.940945] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52993a6a-87aa-7702-3ead-964518d66aea, 'name': SearchDatastore_Task, 'duration_secs': 0.030571} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.940945] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2aa7915a-79d3-4ea7-855b-bda6a948191f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.949935] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 831.949935] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a25f96-2469-270e-dedf-98b0f3693a9d" [ 831.949935] env[61824]: _type = "Task" [ 831.949935] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.960390] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a25f96-2469-270e-dedf-98b0f3693a9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.033044] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275562, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.043373] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94337f08-135b-4fb4-b0d2-3da5df3af692 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.071128] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance 'a33e2573-6d2f-4d07-89d4-f1c6d9d9b382' progress to 0 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 832.150972] env[61824]: DEBUG oslo_concurrency.lockutils [req-02416118-7bff-4dc9-8b46-a0cc797c1141 req-c7a4a6af-8315-49e4-832c-6f3aa53f6ef6 service nova] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.442086] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5282d1e7-c134-46de-b9b1-35412d0df63f tempest-InstanceActionsNegativeTestJSON-1294130836 tempest-InstanceActionsNegativeTestJSON-1294130836-project-member] Lock "0201e783-ed75-4cee-8df4-efac97b4ca19" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.518s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.469387] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a25f96-2469-270e-dedf-98b0f3693a9d, 'name': SearchDatastore_Task, 'duration_secs': 0.01054} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.469387] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.469505] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 5569eb0a-fd42-4e2d-b805-bb40e3264a65/5569eb0a-fd42-4e2d-b805-bb40e3264a65.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.469987] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebf9e9f4-ab37-4366-8845-b8203d1f2690 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.478882] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 832.478882] env[61824]: value = "task-1275563" [ 832.478882] env[61824]: _type = "Task" [ 832.478882] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.489230] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275563, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.535361] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275562, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.579165] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.581130] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ff865e6-ed67-46dd-a780-8cca1dd1a2ac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.586694] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 832.586694] env[61824]: value = "task-1275564" [ 832.586694] env[61824]: _type = "Task" [ 832.586694] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.597568] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.996367] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275563, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48959} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.999894] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 5569eb0a-fd42-4e2d-b805-bb40e3264a65/5569eb0a-fd42-4e2d-b805-bb40e3264a65.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.999894] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.000756] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c8453f5-3aeb-43aa-b82d-52806e5789f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.008043] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 833.008043] env[61824]: value = "task-1275565" [ 833.008043] env[61824]: _type = "Task" [ 833.008043] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.017629] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275565, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.032768] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275562, 'name': CloneVM_Task} progress is 95%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.096463] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275564, 'name': PowerOffVM_Task, 'duration_secs': 0.359975} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.096580] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.096775] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance 'a33e2573-6d2f-4d07-89d4-f1c6d9d9b382' progress to 17 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 833.175647] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc66d622-154e-40a4-aa79-f2b4351ad1de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.183843] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0db5b0-8330-406d-901b-292e045ff946 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.223019] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e184a6f2-a997-4bb1-82d0-4423c7ca73a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.231890] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45b35df-89dd-403e-a810-0c6f2c5d6bca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.248954] env[61824]: DEBUG nova.compute.provider_tree [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.518200] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275565, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.223161} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.518708] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.519415] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8beb1e-7ebe-4549-b94d-fae644817e32 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.546146] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 5569eb0a-fd42-4e2d-b805-bb40e3264a65/5569eb0a-fd42-4e2d-b805-bb40e3264a65.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.546912] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-983d99c2-0dc7-4ab8-90c1-e5f15f886d53 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.565158] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275562, 'name': CloneVM_Task, 'duration_secs': 1.629788} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.565876] env[61824]: INFO nova.virt.vmwareapi.vmops [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Created linked-clone VM from snapshot [ 833.566671] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81278ebc-27b8-4e34-b644-86c5a8de1bd3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.573397] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 833.573397] env[61824]: value = "task-1275566" [ 833.573397] env[61824]: _type = "Task" [ 833.573397] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.581313] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Uploading image 619a0c61-29bb-40c8-bc1c-82d0621b6650 {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 833.588544] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.607588] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.607950] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.608140] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.608325] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.608484] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.608714] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.608973] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.609293] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.609520] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.609737] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.609958] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.617273] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 833.617273] env[61824]: value = "vm-274148" [ 833.617273] env[61824]: _type = "VirtualMachine" [ 833.617273] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 833.617631] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2848166c-a39c-4289-a86f-a880ae12f717 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.628251] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-62a46e0f-f601-456d-8ba6-92a1ecd961a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.635169] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lease: (returnval){ [ 833.635169] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521e78b8-858a-a268-2cff-c38f2f0da55a" [ 833.635169] env[61824]: _type = "HttpNfcLease" [ 833.635169] env[61824]: } obtained for exporting VM: (result){ [ 833.635169] env[61824]: value = "vm-274148" [ 833.635169] env[61824]: _type = "VirtualMachine" [ 833.635169] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 833.635601] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the lease: (returnval){ [ 833.635601] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521e78b8-858a-a268-2cff-c38f2f0da55a" [ 833.635601] env[61824]: _type = "HttpNfcLease" [ 833.635601] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 833.637217] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 833.637217] env[61824]: value = "task-1275567" [ 833.637217] env[61824]: _type = "Task" [ 833.637217] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.654605] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275567, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.654903] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 833.654903] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521e78b8-858a-a268-2cff-c38f2f0da55a" [ 833.654903] env[61824]: _type = "HttpNfcLease" [ 833.654903] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 833.753446] env[61824]: DEBUG nova.scheduler.client.report [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.931853] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "31edad40-a743-4b65-8ef0-615a9ba2a225" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.932335] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.086308] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.145282] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 834.145282] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521e78b8-858a-a268-2cff-c38f2f0da55a" [ 834.145282] env[61824]: _type = "HttpNfcLease" [ 834.145282] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 834.149281] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 834.149281] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521e78b8-858a-a268-2cff-c38f2f0da55a" [ 834.149281] env[61824]: _type = "HttpNfcLease" [ 834.149281] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 834.149281] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f83fab2-e1a5-416c-8789-ddbbd9fc0736 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.157097] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275567, 'name': ReconfigVM_Task, 'duration_secs': 0.270383} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.160829] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance 'a33e2573-6d2f-4d07-89d4-f1c6d9d9b382' progress to 33 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 834.164254] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265b32a-8c8a-304e-b708-ce69e110e5d1/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 834.164340] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265b32a-8c8a-304e-b708-ce69e110e5d1/disk-0.vmdk for reading. {{(pid=61824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 834.258941] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.260121] env[61824]: DEBUG nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 834.262150] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.148s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.262388] env[61824]: DEBUG nova.objects.instance [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lazy-loading 'resources' on Instance uuid 43071be0-eaab-415c-8dcb-948bd9702885 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 834.330694] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4acb7f58-5774-4d67-aff4-05ba8e09d3b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.435734] env[61824]: DEBUG nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 834.586593] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275566, 'name': ReconfigVM_Task, 'duration_secs': 0.672313} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.587092] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 5569eb0a-fd42-4e2d-b805-bb40e3264a65/5569eb0a-fd42-4e2d-b805-bb40e3264a65.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.587649] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-50d7ac87-4753-494e-a435-70f658a49690 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.594720] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 834.594720] env[61824]: value = "task-1275569" [ 834.594720] env[61824]: _type = "Task" [ 834.594720] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.604625] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275569, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.668539] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 834.668853] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 834.669209] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 834.669373] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 834.669590] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 834.669756] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 834.670069] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 834.670267] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 834.670625] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 834.670937] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 834.671253] env[61824]: DEBUG nova.virt.hardware [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.678079] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Reconfiguring VM instance instance-0000003b to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 834.679152] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd5fa5c2-eb05-4d66-a51d-abff999832b6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.699800] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 834.699800] env[61824]: value = "task-1275570" [ 834.699800] env[61824]: _type = "Task" [ 834.699800] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.708451] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275570, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.769065] env[61824]: DEBUG nova.compute.utils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 834.774049] env[61824]: DEBUG nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 834.774687] env[61824]: DEBUG nova.network.neutron [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 834.829706] env[61824]: DEBUG nova.policy [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1eb545ac2cf4f0d9ccddb040b17372b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e756c6817698464a9da391d4dd364946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 834.960736] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.062941] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783df4c3-a728-498a-a7e3-70147f1a1faf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.072350] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f3d8af-8606-4642-9721-94f90d210529 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.109858] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77773e3-d646-47ae-816f-20129fcd2016 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.117992] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275569, 'name': Rename_Task, 'duration_secs': 0.219072} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.120591] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 835.120920] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6df22b51-ec1f-49a1-8922-e1a1ad4680f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.123689] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18be0e7c-6bfe-4a67-8e80-586f49d6b3c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.140579] env[61824]: DEBUG nova.compute.provider_tree [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.144288] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 835.144288] env[61824]: value = "task-1275571" [ 835.144288] env[61824]: _type = "Task" [ 835.144288] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.154042] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275571, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.209395] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275570, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.254735] env[61824]: DEBUG nova.network.neutron [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Successfully created port: 3e5959a6-69ca-4b0a-b953-d6e291034bb2 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 835.274627] env[61824]: DEBUG nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 835.644101] env[61824]: DEBUG nova.scheduler.client.report [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.657738] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275571, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.711593] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275570, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.153340] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.159241] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.764s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.161055] env[61824]: INFO nova.compute.claims [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.166115] env[61824]: DEBUG oslo_vmware.api [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275571, 'name': PowerOnVM_Task, 'duration_secs': 0.637574} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.166115] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 836.166115] env[61824]: INFO nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Took 8.34 seconds to spawn the instance on the hypervisor. [ 836.166115] env[61824]: DEBUG nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.167181] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee139e8-7d9c-41f7-942a-0bcae7cb57e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.184638] env[61824]: INFO nova.scheduler.client.report [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Deleted allocations for instance 43071be0-eaab-415c-8dcb-948bd9702885 [ 836.213495] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275570, 'name': ReconfigVM_Task, 'duration_secs': 1.102656} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.214257] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Reconfigured VM instance instance-0000003b to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 836.215653] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567befa7-228a-409d-8026-d97e0da8fc8d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.240452] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] a33e2573-6d2f-4d07-89d4-f1c6d9d9b382/a33e2573-6d2f-4d07-89d4-f1c6d9d9b382.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.241907] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60126b3d-1521-42c9-8a67-0573f7334cf7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.261680] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 836.261680] env[61824]: value = "task-1275572" [ 836.261680] env[61824]: _type = "Task" [ 836.261680] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.273905] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275572, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.285578] env[61824]: DEBUG nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 836.320538] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.320800] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.320962] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.321817] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.322027] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.322167] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.322388] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.322554] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.322727] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.322895] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.323086] env[61824]: DEBUG nova.virt.hardware [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.323961] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6a2046-6a83-43d8-9ec4-39a9f7802335 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.332771] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-836187a0-617b-44f1-9498-a5afe8f246f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.687323] env[61824]: INFO nova.compute.manager [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Took 35.11 seconds to build instance. [ 836.694068] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2c2a7d2d-9011-42ba-a71f-f6fb2ec12815 tempest-VolumesAdminNegativeTest-822341285 tempest-VolumesAdminNegativeTest-822341285-project-member] Lock "43071be0-eaab-415c-8dcb-948bd9702885" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.020s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.725465] env[61824]: DEBUG nova.compute.manager [req-0f1ade13-c82f-4523-9910-264edc48d38c req-33600ac4-0e8a-4880-bb2b-360050a1d2ac service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Received event network-vif-plugged-3e5959a6-69ca-4b0a-b953-d6e291034bb2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.725913] env[61824]: DEBUG oslo_concurrency.lockutils [req-0f1ade13-c82f-4523-9910-264edc48d38c req-33600ac4-0e8a-4880-bb2b-360050a1d2ac service nova] Acquiring lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.725998] env[61824]: DEBUG oslo_concurrency.lockutils [req-0f1ade13-c82f-4523-9910-264edc48d38c req-33600ac4-0e8a-4880-bb2b-360050a1d2ac service nova] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.726240] env[61824]: DEBUG oslo_concurrency.lockutils [req-0f1ade13-c82f-4523-9910-264edc48d38c req-33600ac4-0e8a-4880-bb2b-360050a1d2ac service nova] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.726520] env[61824]: DEBUG nova.compute.manager [req-0f1ade13-c82f-4523-9910-264edc48d38c req-33600ac4-0e8a-4880-bb2b-360050a1d2ac service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] No waiting events found dispatching network-vif-plugged-3e5959a6-69ca-4b0a-b953-d6e291034bb2 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 836.726746] env[61824]: WARNING nova.compute.manager [req-0f1ade13-c82f-4523-9910-264edc48d38c req-33600ac4-0e8a-4880-bb2b-360050a1d2ac service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Received unexpected event network-vif-plugged-3e5959a6-69ca-4b0a-b953-d6e291034bb2 for instance with vm_state building and task_state spawning. [ 836.772125] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275572, 'name': ReconfigVM_Task, 'duration_secs': 0.393203} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.772711] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Reconfigured VM instance instance-0000003b to attach disk [datastore1] a33e2573-6d2f-4d07-89d4-f1c6d9d9b382/a33e2573-6d2f-4d07-89d4-f1c6d9d9b382.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.773197] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance 'a33e2573-6d2f-4d07-89d4-f1c6d9d9b382' progress to 50 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 836.830919] env[61824]: DEBUG nova.network.neutron [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Successfully updated port: 3e5959a6-69ca-4b0a-b953-d6e291034bb2 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 837.067563] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.067963] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.190225] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b7eb3634-f165-43c6-8fd8-6f253d504624 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.328s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.281072] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d779d6c7-b0d7-4671-a1d3-a830d872dfb5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.307329] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dcbc53b-b764-42db-921f-d87d5d31ea95 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.326528] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance 'a33e2573-6d2f-4d07-89d4-f1c6d9d9b382' progress to 67 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 837.334031] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "refresh_cache-1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.334031] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "refresh_cache-1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.334181] env[61824]: DEBUG nova.network.neutron [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 837.442936] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17e1f7f6-b2ca-4822-8495-c471c126c5e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.451468] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f15b28-4261-4d0e-bca3-b00d30ac28f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.489242] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82be517e-522a-418e-9870-5c574fe8b50f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.497671] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88becb9-3712-4378-b367-4da062245792 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.512034] env[61824]: DEBUG nova.compute.provider_tree [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.577569] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.577569] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 837.577569] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Rebuilding the list of instances to heal {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 837.897919] env[61824]: DEBUG nova.network.neutron [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 837.901815] env[61824]: DEBUG nova.network.neutron [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Port e95afb5b-11f8-4ed4-abb0-a12b48c9c14d binding to destination host cpu-1 is already ACTIVE {{(pid=61824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 838.017840] env[61824]: DEBUG nova.scheduler.client.report [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.084133] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 838.084133] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 838.147813] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.147813] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquired lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.147813] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Forcefully refreshing network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 838.147813] env[61824]: DEBUG nova.objects.instance [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lazy-loading 'info_cache' on Instance uuid 7e73f8b5-a138-4455-a392-9a2b2b860558 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.333509] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "37c405be-62a1-4feb-8cae-ea1c924f4f40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.333975] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.357011] env[61824]: DEBUG nova.network.neutron [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Updating instance_info_cache with network_info: [{"id": "3e5959a6-69ca-4b0a-b953-d6e291034bb2", "address": "fa:16:3e:e0:76:a8", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e5959a6-69", "ovs_interfaceid": "3e5959a6-69ca-4b0a-b953-d6e291034bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.508116] env[61824]: DEBUG nova.compute.manager [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.508420] env[61824]: DEBUG nova.compute.manager [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing instance network info cache due to event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.508740] env[61824]: DEBUG oslo_concurrency.lockutils [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.508982] env[61824]: DEBUG oslo_concurrency.lockutils [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.509251] env[61824]: DEBUG nova.network.neutron [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.529509] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.530148] env[61824]: DEBUG nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 838.535408] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.104s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.536619] env[61824]: INFO nova.compute.claims [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.837092] env[61824]: DEBUG nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.859453] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "refresh_cache-1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.862413] env[61824]: DEBUG nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Instance network_info: |[{"id": "3e5959a6-69ca-4b0a-b953-d6e291034bb2", "address": "fa:16:3e:e0:76:a8", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e5959a6-69", "ovs_interfaceid": "3e5959a6-69ca-4b0a-b953-d6e291034bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 838.862569] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:76:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e5959a6-69ca-4b0a-b953-d6e291034bb2', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.874176] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating folder: Project (e756c6817698464a9da391d4dd364946). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.874176] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db35cb9e-989d-44ba-9c7d-a160ced68fde {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.879893] env[61824]: DEBUG nova.compute.manager [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Received event network-changed-3e5959a6-69ca-4b0a-b953-d6e291034bb2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.880509] env[61824]: DEBUG nova.compute.manager [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Refreshing instance network info cache due to event network-changed-3e5959a6-69ca-4b0a-b953-d6e291034bb2. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.880800] env[61824]: DEBUG oslo_concurrency.lockutils [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] Acquiring lock "refresh_cache-1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.881007] env[61824]: DEBUG oslo_concurrency.lockutils [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] Acquired lock "refresh_cache-1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.881239] env[61824]: DEBUG nova.network.neutron [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Refreshing network info cache for port 3e5959a6-69ca-4b0a-b953-d6e291034bb2 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 838.891635] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created folder: Project (e756c6817698464a9da391d4dd364946) in parent group-v274074. [ 838.891960] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating folder: Instances. Parent ref: group-v274149. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 838.892176] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de24058e-2207-4142-af7b-c403f9f54300 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.902052] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created folder: Instances in parent group-v274149. [ 838.902512] env[61824]: DEBUG oslo.service.loopingcall [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.903012] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.903012] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ceeaa065-01b7-463b-8265-2ee45c98d2dd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.932436] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.932436] env[61824]: value = "task-1275575" [ 838.932436] env[61824]: _type = "Task" [ 838.932436] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.942772] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.943035] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.943217] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.950050] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275575, 'name': CreateVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.054320] env[61824]: DEBUG nova.compute.utils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.055943] env[61824]: DEBUG nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.056135] env[61824]: DEBUG nova.network.neutron [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 839.194394] env[61824]: DEBUG nova.policy [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f8e7019b3c754b079d0df72b0312d6f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc352363e7f94dc6b7a99db8aca565a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.292568] env[61824]: DEBUG nova.network.neutron [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updated VIF entry in instance network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 839.292769] env[61824]: DEBUG nova.network.neutron [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.370299] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.447293] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275575, 'name': CreateVM_Task, 'duration_secs': 0.331772} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.448378] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 839.451631] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.454370] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.454370] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 839.454370] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8ff3739-07e7-4183-9c3b-8e7fdad1dc2a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.458647] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 839.458647] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52586385-1698-63c9-e61f-0c5224d5ed6f" [ 839.458647] env[61824]: _type = "Task" [ 839.458647] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.472748] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52586385-1698-63c9-e61f-0c5224d5ed6f, 'name': SearchDatastore_Task, 'duration_secs': 0.009232} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.473139] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.474390] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.474636] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.474817] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.475319] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.475675] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e5b211c2-0ba6-4e8e-a5d9-e165290e3bb2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.484652] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.484880] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.486224] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7a011c1-69c1-4ad8-93c1-5c2e12576ab7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.492730] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 839.492730] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f28bd8-5ba1-8cc5-adf2-9bc832a77f88" [ 839.492730] env[61824]: _type = "Task" [ 839.492730] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.501372] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f28bd8-5ba1-8cc5-adf2-9bc832a77f88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.560322] env[61824]: DEBUG nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 839.644053] env[61824]: DEBUG nova.network.neutron [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Successfully created port: 45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 839.761267] env[61824]: DEBUG nova.network.neutron [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Updated VIF entry in instance network info cache for port 3e5959a6-69ca-4b0a-b953-d6e291034bb2. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 839.761572] env[61824]: DEBUG nova.network.neutron [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Updating instance_info_cache with network_info: [{"id": "3e5959a6-69ca-4b0a-b953-d6e291034bb2", "address": "fa:16:3e:e0:76:a8", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e5959a6-69", "ovs_interfaceid": "3e5959a6-69ca-4b0a-b953-d6e291034bb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.795539] env[61824]: DEBUG oslo_concurrency.lockutils [req-959107d5-6304-4025-aa9f-3d4c0d943154 req-65af6e05-d4e6-4f7e-8a55-6a4622ad2578 service nova] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.863365] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedfb864-e774-47df-9e3f-c97fa59b3ce7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.872283] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3101ce36-e3dd-4fc2-9fc9-d9dace944ba1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.905446] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e47b10-5929-4b15-88ca-5c67c4939c72 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.919969] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab00bbf4-4adb-49b8-b2da-4b671e94ee82 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.936112] env[61824]: DEBUG nova.compute.provider_tree [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.003531] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f28bd8-5ba1-8cc5-adf2-9bc832a77f88, 'name': SearchDatastore_Task, 'duration_secs': 0.008173} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.004618] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c6b61a9-7e86-4f1f-95d0-e6c17e815475 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.012492] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 840.012492] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52989ac4-d174-8f9a-8999-ae779a73648a" [ 840.012492] env[61824]: _type = "Task" [ 840.012492] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.021624] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52989ac4-d174-8f9a-8999-ae779a73648a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.029510] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.029510] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.029510] env[61824]: DEBUG nova.network.neutron [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 840.048755] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.048755] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.120180] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Updating instance_info_cache with network_info: [{"id": "51fba888-670e-4f85-97a5-0133e0624342", "address": "fa:16:3e:d2:52:4a", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51fba888-67", "ovs_interfaceid": "51fba888-670e-4f85-97a5-0133e0624342", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.270392] env[61824]: DEBUG oslo_concurrency.lockutils [req-c0666903-2471-44b2-924d-894c9e42440b req-6c132a19-d3a9-4da8-a121-023e5acef195 service nova] Releasing lock "refresh_cache-1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.439235] env[61824]: DEBUG nova.scheduler.client.report [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.524331] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52989ac4-d174-8f9a-8999-ae779a73648a, 'name': SearchDatastore_Task, 'duration_secs': 0.015975} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.524331] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.525461] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576/1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.525461] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45379c51-757b-4c68-8fda-e422324f178c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.532489] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 840.532489] env[61824]: value = "task-1275576" [ 840.532489] env[61824]: _type = "Task" [ 840.532489] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.542992] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275576, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.552229] env[61824]: DEBUG nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 840.571205] env[61824]: DEBUG nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 840.599516] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 840.599793] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 840.599920] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 840.600120] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 840.600282] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 840.600455] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 840.600677] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 840.600840] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 840.601012] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 840.601179] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 840.601349] env[61824]: DEBUG nova.virt.hardware [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 840.602540] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2549bd54-5b94-4223-b1f9-c519518e50d8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.616980] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3107978d-b291-4210-83ef-c4f0aec4a21b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.621620] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Releasing lock "refresh_cache-7e73f8b5-a138-4455-a392-9a2b2b860558" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.621822] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Updated the network info_cache for instance {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 840.622023] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.622708] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.622925] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.623131] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.623316] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.623497] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.623627] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 840.623765] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 840.848725] env[61824]: DEBUG nova.network.neutron [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance_info_cache with network_info: [{"id": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "address": "fa:16:3e:ce:e9:8d", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape95afb5b-11", "ovs_interfaceid": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.944077] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.944766] env[61824]: DEBUG nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.949430] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.989s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.956425] env[61824]: INFO nova.compute.claims [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.042737] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275576, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474079} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.043032] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576/1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.043259] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.043516] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dae631b3-68d4-420f-929e-b1772b68384c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.052742] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 841.052742] env[61824]: value = "task-1275577" [ 841.052742] env[61824]: _type = "Task" [ 841.052742] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.060248] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275577, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.082301] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.173078] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.318811] env[61824]: DEBUG nova.compute.manager [req-b1d0a4c7-48ab-4e7a-bef2-d9e44c901172 req-c018f810-46bf-4995-853f-7c762386ef6d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received event network-vif-plugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 841.319149] env[61824]: DEBUG oslo_concurrency.lockutils [req-b1d0a4c7-48ab-4e7a-bef2-d9e44c901172 req-c018f810-46bf-4995-853f-7c762386ef6d service nova] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.319301] env[61824]: DEBUG oslo_concurrency.lockutils [req-b1d0a4c7-48ab-4e7a-bef2-d9e44c901172 req-c018f810-46bf-4995-853f-7c762386ef6d service nova] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.319470] env[61824]: DEBUG oslo_concurrency.lockutils [req-b1d0a4c7-48ab-4e7a-bef2-d9e44c901172 req-c018f810-46bf-4995-853f-7c762386ef6d service nova] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.319645] env[61824]: DEBUG nova.compute.manager [req-b1d0a4c7-48ab-4e7a-bef2-d9e44c901172 req-c018f810-46bf-4995-853f-7c762386ef6d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] No waiting events found dispatching network-vif-plugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 841.319815] env[61824]: WARNING nova.compute.manager [req-b1d0a4c7-48ab-4e7a-bef2-d9e44c901172 req-c018f810-46bf-4995-853f-7c762386ef6d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received unexpected event network-vif-plugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 for instance with vm_state building and task_state spawning. [ 841.352336] env[61824]: DEBUG oslo_concurrency.lockutils [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.425669] env[61824]: DEBUG nova.network.neutron [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Successfully updated port: 45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 841.461946] env[61824]: DEBUG nova.compute.utils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.468055] env[61824]: DEBUG nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 841.468055] env[61824]: DEBUG nova.network.neutron [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 841.524135] env[61824]: DEBUG nova.policy [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff4fff56a7cc4f19b92786ea07b66e8b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bad9de0b032c47429ba879d2d62c6af6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 841.561264] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275577, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069564} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.561552] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.562444] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b955b7-54e2-479f-987d-46a4eecb6f87 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.589974] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576/1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.590053] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95097e18-c5f0-482d-a860-871adb22deb3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.611428] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 841.611428] env[61824]: value = "task-1275578" [ 841.611428] env[61824]: _type = "Task" [ 841.611428] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.622376] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275578, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.886324] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109ce1f4-7694-4e4a-86a1-2ddbf7806a73 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.918030] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265b32a-8c8a-304e-b708-ce69e110e5d1/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 841.918872] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b54bf08-b208-4bc5-a7d5-d1ca8b046127 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.922565] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81123cd8-6bc2-4db3-845c-1e9491649c0e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.926038] env[61824]: DEBUG nova.network.neutron [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Successfully created port: e0fb98c3-2363-41c7-b63f-f3550b0050be {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.928547] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.928683] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.928826] env[61824]: DEBUG nova.network.neutron [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.934398] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265b32a-8c8a-304e-b708-ce69e110e5d1/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 841.934398] env[61824]: ERROR oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265b32a-8c8a-304e-b708-ce69e110e5d1/disk-0.vmdk due to incomplete transfer. [ 841.936750] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8eef234f-bfb7-4170-9db3-d809348c8774 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.938807] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance 'a33e2573-6d2f-4d07-89d4-f1c6d9d9b382' progress to 83 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 841.950713] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265b32a-8c8a-304e-b708-ce69e110e5d1/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 841.951027] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Uploaded image 619a0c61-29bb-40c8-bc1c-82d0621b6650 to the Glance image server {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 841.953275] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Destroying the VM {{(pid=61824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 841.953938] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3ce6844b-1546-47f2-81be-c15a48af9dbb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.961055] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 841.961055] env[61824]: value = "task-1275579" [ 841.961055] env[61824]: _type = "Task" [ 841.961055] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.973091] env[61824]: DEBUG nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.975906] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275579, 'name': Destroy_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.124747] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275578, 'name': ReconfigVM_Task, 'duration_secs': 0.294098} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.125801] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576/1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.128964] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26cf9215-6cfa-4e05-bc4a-b5ea52787bcc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.137745] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 842.137745] env[61824]: value = "task-1275580" [ 842.137745] env[61824]: _type = "Task" [ 842.137745] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.145275] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275580, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.246489] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295d733f-5084-4ae6-84cf-4d20a8cdafb1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.258822] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a5bf502-f8d6-49ea-96c9-c8838d079573 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.300338] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fc85228-5a55-498b-95cc-b37807b2423d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.310409] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b88f3d3-bda6-4d4a-bbf6-5fc15c3906f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.329922] env[61824]: DEBUG nova.compute.provider_tree [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.448017] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.448119] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39576071-0c21-464b-ac84-2be72df855d0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.455549] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 842.455549] env[61824]: value = "task-1275581" [ 842.455549] env[61824]: _type = "Task" [ 842.455549] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.463071] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275581, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.463819] env[61824]: DEBUG nova.network.neutron [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.475143] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275579, 'name': Destroy_Task, 'duration_secs': 0.51189} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.475392] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Destroyed the VM [ 842.475731] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Deleting Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 842.476036] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bf7cddb9-5776-4c58-861a-3f6ad90a0a31 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.484601] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 842.484601] env[61824]: value = "task-1275582" [ 842.484601] env[61824]: _type = "Task" [ 842.484601] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.492960] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275582, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.603136] env[61824]: DEBUG nova.network.neutron [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45dc4201-9f", "ovs_interfaceid": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.647525] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275580, 'name': Rename_Task, 'duration_secs': 0.143691} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.649172] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.649172] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4415abe-e109-4d8b-9aca-544f525f13a2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.656393] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 842.656393] env[61824]: value = "task-1275583" [ 842.656393] env[61824]: _type = "Task" [ 842.656393] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.669835] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.833523] env[61824]: DEBUG nova.scheduler.client.report [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.968425] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275581, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.984726] env[61824]: DEBUG nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.999014] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275582, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.022691] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 843.023133] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 843.023415] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.023742] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 843.024026] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.024306] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 843.024928] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 843.024928] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 843.025293] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 843.025596] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 843.025930] env[61824]: DEBUG nova.virt.hardware [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 843.027309] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8567e4-caae-49c4-bbab-ce5b5d1c0655 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.037331] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc72ef3-b3f3-4743-99f5-3bad50d35f53 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.106063] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.106427] env[61824]: DEBUG nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Instance network_info: |[{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45dc4201-9f", "ovs_interfaceid": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 843.106920] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:8d:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45dc4201-9fb0-4060-ba24-f376e69a6e45', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 843.118027] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating folder: Project (dc352363e7f94dc6b7a99db8aca565a6). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.118027] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7775e1c7-0d9a-444e-ab58-1931835bf00c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.130090] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Created folder: Project (dc352363e7f94dc6b7a99db8aca565a6) in parent group-v274074. [ 843.130326] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating folder: Instances. Parent ref: group-v274152. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 843.130607] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-12137d2c-3f9e-4ca6-9178-756250a324ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.140761] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Created folder: Instances in parent group-v274152. [ 843.141137] env[61824]: DEBUG oslo.service.loopingcall [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.141424] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 843.141740] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3ff52b7-c725-43b7-aaf2-146c8badbb80 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.168604] env[61824]: DEBUG oslo_vmware.api [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275583, 'name': PowerOnVM_Task, 'duration_secs': 0.490032} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.170232] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.171351] env[61824]: INFO nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Took 6.88 seconds to spawn the instance on the hypervisor. [ 843.171867] env[61824]: DEBUG nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.171969] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 843.171969] env[61824]: value = "task-1275586" [ 843.171969] env[61824]: _type = "Task" [ 843.171969] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.172706] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5716f0e2-e422-4861-8ae0-aea25a5d776b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.188563] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275586, 'name': CreateVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.341065] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.341065] env[61824]: DEBUG nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 843.344392] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.974s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.349483] env[61824]: INFO nova.compute.claims [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.358204] env[61824]: DEBUG nova.compute.manager [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received event network-changed-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 843.358204] env[61824]: DEBUG nova.compute.manager [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Refreshing instance network info cache due to event network-changed-45dc4201-9fb0-4060-ba24-f376e69a6e45. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 843.358204] env[61824]: DEBUG oslo_concurrency.lockutils [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] Acquiring lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.358204] env[61824]: DEBUG oslo_concurrency.lockutils [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] Acquired lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.358204] env[61824]: DEBUG nova.network.neutron [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Refreshing network info cache for port 45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 843.471696] env[61824]: DEBUG oslo_vmware.api [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275581, 'name': PowerOnVM_Task, 'duration_secs': 0.583944} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.472046] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.472311] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-caa05206-2766-4c75-920b-b8e47765508a tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance 'a33e2573-6d2f-4d07-89d4-f1c6d9d9b382' progress to 100 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 843.498581] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275582, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.632282] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "5806faad-2bab-4fda-a131-566af50c4d7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.632510] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5806faad-2bab-4fda-a131-566af50c4d7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.685968] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275586, 'name': CreateVM_Task, 'duration_secs': 0.434326} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.686224] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.687022] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.687264] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.687647] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.687949] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e410edd-2102-4bf0-95ca-1262f1c5b28c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.698866] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 843.698866] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52edf852-eef0-735e-f267-076dcc05d868" [ 843.698866] env[61824]: _type = "Task" [ 843.698866] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.698866] env[61824]: INFO nova.compute.manager [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Took 22.17 seconds to build instance. [ 843.708578] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52edf852-eef0-735e-f267-076dcc05d868, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.753494] env[61824]: DEBUG nova.network.neutron [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Successfully updated port: e0fb98c3-2363-41c7-b63f-f3550b0050be {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.859048] env[61824]: DEBUG nova.compute.utils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.859724] env[61824]: DEBUG nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.859908] env[61824]: DEBUG nova.network.neutron [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 843.947386] env[61824]: DEBUG nova.policy [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3f6b6ab0258348f3b9930abe3cd34ce8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '869f8aef6de3495895edecbc56abe852', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 844.001268] env[61824]: DEBUG oslo_vmware.api [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275582, 'name': RemoveSnapshot_Task, 'duration_secs': 1.041241} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.001628] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Deleted Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 844.003241] env[61824]: INFO nova.compute.manager [None req-9d0ff459-e4c9-4f5f-ac15-ab05686a19d6 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Took 15.08 seconds to snapshot the instance on the hypervisor. [ 844.135040] env[61824]: DEBUG nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 844.204248] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f274bf0-cba7-4c38-9e21-89b77e86622a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.688s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.212137] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52edf852-eef0-735e-f267-076dcc05d868, 'name': SearchDatastore_Task, 'duration_secs': 0.012621} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.212137] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.212137] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.212137] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.212330] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.212414] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.212680] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4d69a85-80fc-4fa4-a0c6-a5939922a172 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.230453] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.230645] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.231409] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4d1d78b-9dbb-4f6a-8856-e8d36c3bafea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.237197] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 844.237197] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6b2dd-ad9a-46fe-41d6-c01c83e1d9c9" [ 844.237197] env[61824]: _type = "Task" [ 844.237197] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.246211] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6b2dd-ad9a-46fe-41d6-c01c83e1d9c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.257443] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.257443] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.257443] env[61824]: DEBUG nova.network.neutron [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.296367] env[61824]: DEBUG nova.network.neutron [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updated VIF entry in instance network info cache for port 45dc4201-9fb0-4060-ba24-f376e69a6e45. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 844.296457] env[61824]: DEBUG nova.network.neutron [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45dc4201-9f", "ovs_interfaceid": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.363816] env[61824]: DEBUG nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 844.403134] env[61824]: DEBUG nova.network.neutron [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Successfully created port: 8ff3d743-0e3d-47f7-85ad-8981881e11ac {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.642750] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0b4f52-56e0-461b-813c-6c2d8e3ffb1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.652132] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cddf9e-b11e-4444-9dcc-83d6de92716c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.656508] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.684295] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040efd36-c0d2-4b32-aba9-ffdb75ce09b4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.691739] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d99d187-346a-455e-b81e-c3d9bb9b19ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.704771] env[61824]: DEBUG nova.compute.provider_tree [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.748106] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6b2dd-ad9a-46fe-41d6-c01c83e1d9c9, 'name': SearchDatastore_Task, 'duration_secs': 0.024951} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.748911] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c48feb94-cd2a-4d66-8c2d-066b10add369 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.753810] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 844.753810] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52297399-c72b-6feb-0274-3ea1d7aa613e" [ 844.753810] env[61824]: _type = "Task" [ 844.753810] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.762699] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52297399-c72b-6feb-0274-3ea1d7aa613e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.800281] env[61824]: DEBUG oslo_concurrency.lockutils [req-134414a7-ca58-4d67-b4a6-5535240216db req-e333409c-1c48-4d0b-b339-2b203ca13b3d service nova] Releasing lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.803340] env[61824]: DEBUG nova.network.neutron [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.207836] env[61824]: DEBUG nova.scheduler.client.report [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.265800] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52297399-c72b-6feb-0274-3ea1d7aa613e, 'name': SearchDatastore_Task, 'duration_secs': 0.022909} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.265800] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.266029] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.266264] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58d208fa-93ab-4810-991b-cb092f6b5714 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.273976] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 845.273976] env[61824]: value = "task-1275587" [ 845.273976] env[61824]: _type = "Task" [ 845.273976] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.282933] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275587, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.284139] env[61824]: DEBUG nova.network.neutron [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Updating instance_info_cache with network_info: [{"id": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "address": "fa:16:3e:a3:4c:22", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0fb98c3-23", "ovs_interfaceid": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.373101] env[61824]: DEBUG nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 845.395015] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 845.395015] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 845.395264] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 845.395350] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 845.395589] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 845.395741] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 845.396022] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 845.396238] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 845.396483] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 845.396625] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 845.396889] env[61824]: DEBUG nova.virt.hardware [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 845.397893] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbbe10d-961b-4c75-bf86-062cc08065bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.407408] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d37e8a-67d9-46a8-b57c-3b0f99dbbaca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.584258] env[61824]: DEBUG nova.compute.manager [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Received event network-vif-plugged-e0fb98c3-2363-41c7-b63f-f3550b0050be {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.584393] env[61824]: DEBUG oslo_concurrency.lockutils [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] Acquiring lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.584621] env[61824]: DEBUG oslo_concurrency.lockutils [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.584768] env[61824]: DEBUG oslo_concurrency.lockutils [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.584941] env[61824]: DEBUG nova.compute.manager [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] No waiting events found dispatching network-vif-plugged-e0fb98c3-2363-41c7-b63f-f3550b0050be {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 845.585127] env[61824]: WARNING nova.compute.manager [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Received unexpected event network-vif-plugged-e0fb98c3-2363-41c7-b63f-f3550b0050be for instance with vm_state building and task_state spawning. [ 845.585295] env[61824]: DEBUG nova.compute.manager [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Received event network-changed-e0fb98c3-2363-41c7-b63f-f3550b0050be {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.585453] env[61824]: DEBUG nova.compute.manager [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Refreshing instance network info cache due to event network-changed-e0fb98c3-2363-41c7-b63f-f3550b0050be. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.585619] env[61824]: DEBUG oslo_concurrency.lockutils [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] Acquiring lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.713010] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.713677] env[61824]: DEBUG nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.716414] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.634s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.718249] env[61824]: INFO nova.compute.claims [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.785123] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275587, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.786746] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.787098] env[61824]: DEBUG nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Instance network_info: |[{"id": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "address": "fa:16:3e:a3:4c:22", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0fb98c3-23", "ovs_interfaceid": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.787411] env[61824]: DEBUG oslo_concurrency.lockutils [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] Acquired lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.787602] env[61824]: DEBUG nova.network.neutron [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Refreshing network info cache for port e0fb98c3-2363-41c7-b63f-f3550b0050be {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 845.788955] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:4c:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6076d24d-3c8e-4bbb-ba96-a08fb27a73cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0fb98c3-2363-41c7-b63f-f3550b0050be', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.799419] env[61824]: DEBUG oslo.service.loopingcall [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.800587] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.800886] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8769a7e0-5cec-48b7-a76f-c9329bc03af4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.830269] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.830269] env[61824]: value = "task-1275588" [ 845.830269] env[61824]: _type = "Task" [ 845.830269] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.844877] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275588, 'name': CreateVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.094395] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.094788] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.095254] env[61824]: DEBUG nova.compute.manager [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Going to confirm migration 1 {{(pid=61824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 846.225157] env[61824]: DEBUG nova.compute.utils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.227914] env[61824]: DEBUG nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 846.229066] env[61824]: DEBUG nova.network.neutron [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 846.285944] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275587, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.713482} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.286303] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.286757] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.287082] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86f4a348-ed42-4ceb-8e32-25ae06bcf116 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.293978] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 846.293978] env[61824]: value = "task-1275589" [ 846.293978] env[61824]: _type = "Task" [ 846.293978] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.307726] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275589, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.316098] env[61824]: DEBUG nova.policy [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '46b60264e47743d08c7dcc06e7b231ae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8b4f2f30c68b44cf8ae59b116d98f6c7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 846.344379] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275588, 'name': CreateVM_Task, 'duration_secs': 0.428161} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.344724] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.345279] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.345453] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.345776] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.346072] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10018161-9ffa-47e7-9f89-d07888802eb4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.353129] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 846.353129] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5231f751-eed7-5352-3bb1-1a6522245d3f" [ 846.353129] env[61824]: _type = "Task" [ 846.353129] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.364372] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231f751-eed7-5352-3bb1-1a6522245d3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.539407] env[61824]: DEBUG nova.network.neutron [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Updated VIF entry in instance network info cache for port e0fb98c3-2363-41c7-b63f-f3550b0050be. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.540860] env[61824]: DEBUG nova.network.neutron [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Updating instance_info_cache with network_info: [{"id": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "address": "fa:16:3e:a3:4c:22", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0fb98c3-23", "ovs_interfaceid": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.647314] env[61824]: DEBUG nova.network.neutron [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Successfully updated port: 8ff3d743-0e3d-47f7-85ad-8981881e11ac {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 846.699301] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.699499] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.699678] env[61824]: DEBUG nova.network.neutron [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 846.699869] env[61824]: DEBUG nova.objects.instance [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lazy-loading 'info_cache' on Instance uuid a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.729165] env[61824]: DEBUG nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.734856] env[61824]: DEBUG nova.network.neutron [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Successfully created port: beb8cba8-71b3-40a5-ba5d-665a9770ec4d {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.810856] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275589, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.204157} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.811959] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.813611] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6e4e96-24f6-4482-85a0-febaaa469107 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.845548] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.849869] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9f4949f-3a45-423c-8b27-f1592e68f3f4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.867928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "cabad7a6-206f-4e02-be3c-e18654396b7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.867928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.883190] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 846.883190] env[61824]: value = "task-1275590" [ 846.883190] env[61824]: _type = "Task" [ 846.883190] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.883493] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231f751-eed7-5352-3bb1-1a6522245d3f, 'name': SearchDatastore_Task, 'duration_secs': 0.021545} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.883896] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.884208] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.884513] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.884692] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.884909] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.888321] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9677a3d-5b9a-41a7-aa47-f1accd007bff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.899602] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.901730] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.901922] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.902815] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0059340-ba3a-40cb-9af3-5b9a7b34bce8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.910848] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 846.910848] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f1b2fa-c729-06e6-3835-662514f08424" [ 846.910848] env[61824]: _type = "Task" [ 846.910848] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.918987] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f1b2fa-c729-06e6-3835-662514f08424, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.043187] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ce772a-7d4b-4985-9dbe-38ac66ad3f6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.048318] env[61824]: DEBUG oslo_concurrency.lockutils [req-c6b74f1c-6c60-49a1-ac35-891936c41dab req-258edf53-f8c6-41b1-99ed-c0cb58aeb696 service nova] Releasing lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.053467] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be621f13-4cf6-485f-b7d9-26311f134f4d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.083475] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7071a43c-3571-45d8-9747-3bcb667e079c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.091303] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17034a19-79ce-4cd8-9d01-1ea64d67ecaf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.104724] env[61824]: DEBUG nova.compute.provider_tree [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.151805] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "refresh_cache-31edad40-a743-4b65-8ef0-615a9ba2a225" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.151991] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquired lock "refresh_cache-31edad40-a743-4b65-8ef0-615a9ba2a225" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.152197] env[61824]: DEBUG nova.network.neutron [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 847.190557] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Acquiring lock "6235a9b9-19c0-4917-938a-bca82a2ca05e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.190821] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.191058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Acquiring lock "6235a9b9-19c0-4917-938a-bca82a2ca05e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.191329] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.194217] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.194217] env[61824]: INFO nova.compute.manager [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Terminating instance [ 847.196107] env[61824]: DEBUG nova.compute.manager [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 847.196328] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 847.196560] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c3736e9-6eeb-4efb-a9a4-381a18c2ff1b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.210623] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 847.210623] env[61824]: value = "task-1275591" [ 847.210623] env[61824]: _type = "Task" [ 847.210623] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.220036] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.373854] env[61824]: DEBUG nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.394067] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275590, 'name': ReconfigVM_Task, 'duration_secs': 0.287174} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.394481] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.395156] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfe9a508-1273-4087-89c9-ebfe7273580e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.403419] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 847.403419] env[61824]: value = "task-1275592" [ 847.403419] env[61824]: _type = "Task" [ 847.403419] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.413815] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275592, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.422889] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f1b2fa-c729-06e6-3835-662514f08424, 'name': SearchDatastore_Task, 'duration_secs': 0.027381} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.423623] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ae11ee6-68dd-406a-9486-850627c36f6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.428814] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 847.428814] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a297f1-b17f-3bf5-f44b-21e8cfd8127c" [ 847.428814] env[61824]: _type = "Task" [ 847.428814] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.436986] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a297f1-b17f-3bf5-f44b-21e8cfd8127c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.608315] env[61824]: DEBUG nova.scheduler.client.report [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.613817] env[61824]: DEBUG nova.compute.manager [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Received event network-vif-plugged-8ff3d743-0e3d-47f7-85ad-8981881e11ac {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.614116] env[61824]: DEBUG oslo_concurrency.lockutils [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] Acquiring lock "31edad40-a743-4b65-8ef0-615a9ba2a225-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.614396] env[61824]: DEBUG oslo_concurrency.lockutils [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.614629] env[61824]: DEBUG oslo_concurrency.lockutils [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.614857] env[61824]: DEBUG nova.compute.manager [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] No waiting events found dispatching network-vif-plugged-8ff3d743-0e3d-47f7-85ad-8981881e11ac {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.615109] env[61824]: WARNING nova.compute.manager [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Received unexpected event network-vif-plugged-8ff3d743-0e3d-47f7-85ad-8981881e11ac for instance with vm_state building and task_state spawning. [ 847.615342] env[61824]: DEBUG nova.compute.manager [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Received event network-changed-8ff3d743-0e3d-47f7-85ad-8981881e11ac {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.615564] env[61824]: DEBUG nova.compute.manager [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Refreshing instance network info cache due to event network-changed-8ff3d743-0e3d-47f7-85ad-8981881e11ac. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.615802] env[61824]: DEBUG oslo_concurrency.lockutils [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] Acquiring lock "refresh_cache-31edad40-a743-4b65-8ef0-615a9ba2a225" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.684272] env[61824]: DEBUG nova.network.neutron [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 847.722706] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275591, 'name': PowerOffVM_Task, 'duration_secs': 0.334285} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.724741] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 847.724955] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 847.725233] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274093', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'name': 'volume-32104921-3bcd-4691-86e7-9c3c0a08ed72', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6235a9b9-19c0-4917-938a-bca82a2ca05e', 'attached_at': '', 'detached_at': '', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'serial': '32104921-3bcd-4691-86e7-9c3c0a08ed72'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 847.725946] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8c8bc0-2bde-4c31-a7de-b90932dfd090 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.745399] env[61824]: DEBUG nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.751182] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc6c0acc-53bb-4380-b738-37a861eb452a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.760655] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26cc4a20-eb62-4ce7-82a9-4c7c7e70a888 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.780123] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.780360] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.780518] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.780707] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.780889] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.781078] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.781286] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.781451] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.781618] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.781803] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.781957] env[61824]: DEBUG nova.virt.hardware [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.782750] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45bf1b0-8665-4aaa-bd9d-5cbc1f00d4ba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.785802] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d894afe-a54b-41c6-be79-5bfed8b42744 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.804286] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] The volume has not been displaced from its original location: [datastore2] volume-32104921-3bcd-4691-86e7-9c3c0a08ed72/volume-32104921-3bcd-4691-86e7-9c3c0a08ed72.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 847.809769] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Reconfiguring VM instance instance-0000003d to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 847.813409] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8abf6f-1e6e-4791-acab-22015264ddb1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.818912] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a597839-fdd3-4b74-bc47-cd1186a288f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.847210] env[61824]: DEBUG nova.network.neutron [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Updating instance_info_cache with network_info: [{"id": "8ff3d743-0e3d-47f7-85ad-8981881e11ac", "address": "fa:16:3e:9b:23:34", "network": {"id": "62645edc-7636-4e09-9c9f-ea8ac8bfda5c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-953834292-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869f8aef6de3495895edecbc56abe852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c42bb08a-77b4-4bba-8166-702cbb1b5f1e", "external-id": "nsx-vlan-transportzone-137", "segmentation_id": 137, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ff3d743-0e", "ovs_interfaceid": "8ff3d743-0e3d-47f7-85ad-8981881e11ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.848526] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 847.848526] env[61824]: value = "task-1275593" [ 847.848526] env[61824]: _type = "Task" [ 847.848526] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.860476] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275593, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.897670] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.914998] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275592, 'name': Rename_Task, 'duration_secs': 0.322339} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.914998] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.914998] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27f703e9-bd72-4bcf-8021-11e59662af5d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.921791] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 847.921791] env[61824]: value = "task-1275594" [ 847.921791] env[61824]: _type = "Task" [ 847.921791] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.929823] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275594, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.938619] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a297f1-b17f-3bf5-f44b-21e8cfd8127c, 'name': SearchDatastore_Task, 'duration_secs': 0.019579} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.938902] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.939203] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 01fd9faa-a0de-45b2-89bf-094c6d1ac155/01fd9faa-a0de-45b2-89bf-094c6d1ac155.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.939476] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aa391b5c-e7a8-4fe1-8dc3-b0af500af55e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.945804] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 847.945804] env[61824]: value = "task-1275595" [ 847.945804] env[61824]: _type = "Task" [ 847.945804] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.956652] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275595, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.105022] env[61824]: DEBUG nova.network.neutron [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance_info_cache with network_info: [{"id": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "address": "fa:16:3e:ce:e9:8d", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape95afb5b-11", "ovs_interfaceid": "e95afb5b-11f8-4ed4-abb0-a12b48c9c14d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.117558] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.118247] env[61824]: DEBUG nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 848.121582] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.985s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.121803] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.121975] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 848.122352] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.466s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.124467] env[61824]: INFO nova.compute.claims [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.128540] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bde26d0-8f3a-4f3d-bc7a-1bb374ef55e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.138403] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ec12fe-d765-4c23-9864-2ecbd223352d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.153585] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96262b19-5511-4af8-b09c-e82ac0c2d716 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.161435] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e3796f2-0f36-441e-8022-62a26cfe08d2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.197335] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180901MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 848.197528] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.349808] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Releasing lock "refresh_cache-31edad40-a743-4b65-8ef0-615a9ba2a225" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.351394] env[61824]: DEBUG nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Instance network_info: |[{"id": "8ff3d743-0e3d-47f7-85ad-8981881e11ac", "address": "fa:16:3e:9b:23:34", "network": {"id": "62645edc-7636-4e09-9c9f-ea8ac8bfda5c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-953834292-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869f8aef6de3495895edecbc56abe852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c42bb08a-77b4-4bba-8166-702cbb1b5f1e", "external-id": "nsx-vlan-transportzone-137", "segmentation_id": 137, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ff3d743-0e", "ovs_interfaceid": "8ff3d743-0e3d-47f7-85ad-8981881e11ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 848.351394] env[61824]: DEBUG oslo_concurrency.lockutils [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] Acquired lock "refresh_cache-31edad40-a743-4b65-8ef0-615a9ba2a225" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.351546] env[61824]: DEBUG nova.network.neutron [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Refreshing network info cache for port 8ff3d743-0e3d-47f7-85ad-8981881e11ac {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 848.354430] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:23:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c42bb08a-77b4-4bba-8166-702cbb1b5f1e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ff3d743-0e3d-47f7-85ad-8981881e11ac', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 848.373634] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Creating folder: Project (869f8aef6de3495895edecbc56abe852). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.389779] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd332de7-788f-41d2-a387-0b6fadec60cf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.406875] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275593, 'name': ReconfigVM_Task, 'duration_secs': 0.185643} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.407387] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Reconfigured VM instance instance-0000003d to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 848.417225] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cb6ec54-571b-4baf-96b2-eecf17c499d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.428228] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Created folder: Project (869f8aef6de3495895edecbc56abe852) in parent group-v274074. [ 848.428546] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Creating folder: Instances. Parent ref: group-v274156. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 848.429280] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21ce213e-fb5b-4560-8dc5-14111fb1a254 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.442166] env[61824]: DEBUG oslo_vmware.api [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275594, 'name': PowerOnVM_Task, 'duration_secs': 0.5129} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.443964] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.444246] env[61824]: INFO nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Took 7.87 seconds to spawn the instance on the hypervisor. [ 848.444521] env[61824]: DEBUG nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.444906] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 848.444906] env[61824]: value = "task-1275597" [ 848.444906] env[61824]: _type = "Task" [ 848.444906] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.446960] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49a4300-39f9-4f25-b986-b22eac9b8bc9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.451150] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Created folder: Instances in parent group-v274156. [ 848.451495] env[61824]: DEBUG oslo.service.loopingcall [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 848.455910] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 848.459661] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c15fa1e5-16f8-4b08-947e-f65e99992018 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.485634] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.492140] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275595, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532967} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.493163] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 01fd9faa-a0de-45b2-89bf-094c6d1ac155/01fd9faa-a0de-45b2-89bf-094c6d1ac155.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.493163] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.496503] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-670a8e01-a521-4944-a0ca-e775effa233e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.497041] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 848.497041] env[61824]: value = "task-1275599" [ 848.497041] env[61824]: _type = "Task" [ 848.497041] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.501584] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 848.501584] env[61824]: value = "task-1275600" [ 848.501584] env[61824]: _type = "Task" [ 848.501584] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.507851] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275599, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.515558] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275600, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.607561] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.607959] env[61824]: DEBUG nova.objects.instance [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lazy-loading 'migration_context' on Instance uuid a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.629105] env[61824]: DEBUG nova.compute.utils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.630467] env[61824]: DEBUG nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.630636] env[61824]: DEBUG nova.network.neutron [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 848.720685] env[61824]: DEBUG nova.policy [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '316c089785a44aa897040f3c0ef7a071', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0924bfa41b5e42d68818557d2b0ace7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.750301] env[61824]: DEBUG nova.network.neutron [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Successfully updated port: beb8cba8-71b3-40a5-ba5d-665a9770ec4d {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.960015] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275597, 'name': ReconfigVM_Task, 'duration_secs': 0.163404} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.960414] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274093', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'name': 'volume-32104921-3bcd-4691-86e7-9c3c0a08ed72', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6235a9b9-19c0-4917-938a-bca82a2ca05e', 'attached_at': '', 'detached_at': '', 'volume_id': '32104921-3bcd-4691-86e7-9c3c0a08ed72', 'serial': '32104921-3bcd-4691-86e7-9c3c0a08ed72'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 848.960702] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 848.961512] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada8c904-8f3c-46c1-a84c-a6cf1614c89e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.968848] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 848.969013] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6f66885-ca38-49ff-b073-7bf015003368 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.994855] env[61824]: INFO nova.compute.manager [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Took 21.62 seconds to build instance. [ 849.010058] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275599, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.015692] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275600, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074715} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.016549] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.017545] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc42de76-17d5-4991-a825-21c72edcf20a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.044625] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 01fd9faa-a0de-45b2-89bf-094c6d1ac155/01fd9faa-a0de-45b2-89bf-094c6d1ac155.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.048762] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28c301a2-fcf5-4896-823b-587b5fb91781 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.063986] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 849.064270] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 849.064460] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Deleting the datastore file [datastore2] 6235a9b9-19c0-4917-938a-bca82a2ca05e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.065345] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-702f81e1-b6a4-4ff8-ab04-27f9eaa7575e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.072192] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 849.072192] env[61824]: value = "task-1275602" [ 849.072192] env[61824]: _type = "Task" [ 849.072192] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.073448] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for the task: (returnval){ [ 849.073448] env[61824]: value = "task-1275603" [ 849.073448] env[61824]: _type = "Task" [ 849.073448] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.084495] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275602, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.087528] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.110297] env[61824]: DEBUG nova.objects.base [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 849.111226] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b243636-8e7e-43ad-9761-88dbed5b63bb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.130752] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8f69a3e-b468-432c-9042-98fd19580be0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.133280] env[61824]: DEBUG nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.143368] env[61824]: DEBUG oslo_vmware.api [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 849.143368] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5231d3fe-a0f0-90aa-362b-641f4d6a62dc" [ 849.143368] env[61824]: _type = "Task" [ 849.143368] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.152090] env[61824]: DEBUG oslo_vmware.api [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231d3fe-a0f0-90aa-362b-641f4d6a62dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.253309] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "refresh_cache-37c405be-62a1-4feb-8cae-ea1c924f4f40" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.253585] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquired lock "refresh_cache-37c405be-62a1-4feb-8cae-ea1c924f4f40" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.253754] env[61824]: DEBUG nova.network.neutron [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 849.434609] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e96d9c4e-7c88-4b82-86b4-606fea8a7f91 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.449678] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e7e315-33cd-461f-94a5-062596185423 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.483181] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f20efe-69da-4706-9717-a5124d7340ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.491064] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f518f7-f729-4f40-b8dc-3dd13a1b7447 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.506160] env[61824]: DEBUG oslo_concurrency.lockutils [None req-955be206-4b77-4032-8939-29bcb8232fce tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.144s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.506646] env[61824]: DEBUG nova.compute.provider_tree [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 849.518018] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275599, 'name': CreateVM_Task, 'duration_secs': 0.59756} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.518018] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.519860] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.519860] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.519860] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 849.519860] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d25c765c-27be-4907-b4e5-0648da316629 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.526397] env[61824]: DEBUG nova.network.neutron [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Updated VIF entry in instance network info cache for port 8ff3d743-0e3d-47f7-85ad-8981881e11ac. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 849.526706] env[61824]: DEBUG nova.network.neutron [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Updating instance_info_cache with network_info: [{"id": "8ff3d743-0e3d-47f7-85ad-8981881e11ac", "address": "fa:16:3e:9b:23:34", "network": {"id": "62645edc-7636-4e09-9c9f-ea8ac8bfda5c", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-953834292-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869f8aef6de3495895edecbc56abe852", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c42bb08a-77b4-4bba-8166-702cbb1b5f1e", "external-id": "nsx-vlan-transportzone-137", "segmentation_id": 137, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ff3d743-0e", "ovs_interfaceid": "8ff3d743-0e3d-47f7-85ad-8981881e11ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.529563] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 849.529563] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52619cc3-3a47-756a-4822-b9da5ea137c9" [ 849.529563] env[61824]: _type = "Task" [ 849.529563] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.531990] env[61824]: DEBUG nova.network.neutron [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Successfully created port: bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.544259] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52619cc3-3a47-756a-4822-b9da5ea137c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.587647] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275602, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.592181] env[61824]: DEBUG oslo_vmware.api [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Task: {'id': task-1275603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.430836} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.592181] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.592181] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 849.592181] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 849.592181] env[61824]: INFO nova.compute.manager [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Took 2.40 seconds to destroy the instance on the hypervisor. [ 849.592396] env[61824]: DEBUG oslo.service.loopingcall [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.592585] env[61824]: DEBUG nova.compute.manager [-] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.592674] env[61824]: DEBUG nova.network.neutron [-] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.657215] env[61824]: DEBUG oslo_vmware.api [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231d3fe-a0f0-90aa-362b-641f4d6a62dc, 'name': SearchDatastore_Task, 'duration_secs': 0.012151} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.657694] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.840508] env[61824]: DEBUG nova.network.neutron [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 850.030569] env[61824]: DEBUG oslo_concurrency.lockutils [req-685dad4b-26c6-4736-83be-908a56c80212 req-3d1cdf93-0e29-4fcd-8513-b4beffb7e36e service nova] Releasing lock "refresh_cache-31edad40-a743-4b65-8ef0-615a9ba2a225" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.043229] env[61824]: ERROR nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [req-ac02fc9a-e90b-41f5-bed6-3f6c7f70b9d2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8d9ece0e-729c-4d69-b741-d89969200860. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ac02fc9a-e90b-41f5-bed6-3f6c7f70b9d2"}]} [ 850.053225] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52619cc3-3a47-756a-4822-b9da5ea137c9, 'name': SearchDatastore_Task, 'duration_secs': 0.029693} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.053959] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.053959] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.054935] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.054935] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.054935] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.054935] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6788f7c5-d6a7-4186-ba4c-702a4dcc67c7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.062273] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing inventories for resource provider 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 850.067874] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.067874] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.067874] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed2400cd-1edf-4851-98d2-56d001d1260c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.073850] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 850.073850] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52177687-ce60-2a7b-5249-3003fa1d0bcf" [ 850.073850] env[61824]: _type = "Task" [ 850.073850] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.084463] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating ProviderTree inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 850.084718] env[61824]: DEBUG nova.compute.provider_tree [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.096439] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52177687-ce60-2a7b-5249-3003fa1d0bcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.096810] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275602, 'name': ReconfigVM_Task, 'duration_secs': 0.928175} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.097047] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 01fd9faa-a0de-45b2-89bf-094c6d1ac155/01fd9faa-a0de-45b2-89bf-094c6d1ac155.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.099672] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-207d53cd-39d0-44e9-a7f7-ee59a24b5aa5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.102898] env[61824]: DEBUG nova.compute.manager [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Received event network-vif-plugged-beb8cba8-71b3-40a5-ba5d-665a9770ec4d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.103197] env[61824]: DEBUG oslo_concurrency.lockutils [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] Acquiring lock "37c405be-62a1-4feb-8cae-ea1c924f4f40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.103389] env[61824]: DEBUG oslo_concurrency.lockutils [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.103580] env[61824]: DEBUG oslo_concurrency.lockutils [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.103752] env[61824]: DEBUG nova.compute.manager [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] No waiting events found dispatching network-vif-plugged-beb8cba8-71b3-40a5-ba5d-665a9770ec4d {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.103923] env[61824]: WARNING nova.compute.manager [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Received unexpected event network-vif-plugged-beb8cba8-71b3-40a5-ba5d-665a9770ec4d for instance with vm_state building and task_state spawning. [ 850.104314] env[61824]: DEBUG nova.compute.manager [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Received event network-changed-beb8cba8-71b3-40a5-ba5d-665a9770ec4d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.104525] env[61824]: DEBUG nova.compute.manager [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Refreshing instance network info cache due to event network-changed-beb8cba8-71b3-40a5-ba5d-665a9770ec4d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.104706] env[61824]: DEBUG oslo_concurrency.lockutils [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] Acquiring lock "refresh_cache-37c405be-62a1-4feb-8cae-ea1c924f4f40" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.105791] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing aggregate associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, aggregates: None {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 850.115149] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 850.115149] env[61824]: value = "task-1275604" [ 850.115149] env[61824]: _type = "Task" [ 850.115149] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.128368] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275604, 'name': Rename_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.129451] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing trait associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 850.145844] env[61824]: DEBUG nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 850.181297] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.181297] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.181297] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.181459] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.181459] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.181459] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.181459] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.181459] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.181598] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.181598] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.181598] env[61824]: DEBUG nova.virt.hardware [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.182432] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6245fd6f-9af6-47af-81b2-39a225b989d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.193948] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f4f09f-9198-484f-b3bd-5b2ade141dcf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.387456] env[61824]: DEBUG nova.network.neutron [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Updating instance_info_cache with network_info: [{"id": "beb8cba8-71b3-40a5-ba5d-665a9770ec4d", "address": "fa:16:3e:75:25:03", "network": {"id": "9f8b7344-1620-45fe-ac3c-2d195c378f3e", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-883178807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b4f2f30c68b44cf8ae59b116d98f6c7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb8cba8-71", "ovs_interfaceid": "beb8cba8-71b3-40a5-ba5d-665a9770ec4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.444572] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d47592e-a016-446e-ab16-89ac7384ed2b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.454764] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b661c23-a927-46d3-8af3-01d3ddcc702d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.490623] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37babe86-881a-4a04-bb97-a8dbe8f01dfe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.498758] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8538eaba-bb2d-4278-bee4-ba182cb08a68 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.513757] env[61824]: DEBUG nova.compute.provider_tree [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.591336] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52177687-ce60-2a7b-5249-3003fa1d0bcf, 'name': SearchDatastore_Task, 'duration_secs': 0.020775} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.592756] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b098d6ce-5389-498e-82e9-7fcda25d4d81 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.601299] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 850.601299] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52348a35-564e-a4ca-91a1-530481f01e3a" [ 850.601299] env[61824]: _type = "Task" [ 850.601299] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.615243] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52348a35-564e-a4ca-91a1-530481f01e3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.630625] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275604, 'name': Rename_Task, 'duration_secs': 0.238944} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.631188] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.631629] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c08fed1-18c6-44e9-b767-6c7435c2a4ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.639415] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 850.639415] env[61824]: value = "task-1275605" [ 850.639415] env[61824]: _type = "Task" [ 850.639415] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.649073] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275605, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.729147] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "c6552176-1128-473b-ac4a-f3b55428fd9e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.729467] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.890906] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Releasing lock "refresh_cache-37c405be-62a1-4feb-8cae-ea1c924f4f40" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.891296] env[61824]: DEBUG nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Instance network_info: |[{"id": "beb8cba8-71b3-40a5-ba5d-665a9770ec4d", "address": "fa:16:3e:75:25:03", "network": {"id": "9f8b7344-1620-45fe-ac3c-2d195c378f3e", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-883178807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b4f2f30c68b44cf8ae59b116d98f6c7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb8cba8-71", "ovs_interfaceid": "beb8cba8-71b3-40a5-ba5d-665a9770ec4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 850.891627] env[61824]: DEBUG oslo_concurrency.lockutils [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] Acquired lock "refresh_cache-37c405be-62a1-4feb-8cae-ea1c924f4f40" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.891813] env[61824]: DEBUG nova.network.neutron [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Refreshing network info cache for port beb8cba8-71b3-40a5-ba5d-665a9770ec4d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 850.893083] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:25:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49a4d142-3f97-47fe-b074-58923c46815e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'beb8cba8-71b3-40a5-ba5d-665a9770ec4d', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.902288] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Creating folder: Project (8b4f2f30c68b44cf8ae59b116d98f6c7). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.903559] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a143ba44-7847-4eb1-b685-76b37c99cead {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.915080] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Created folder: Project (8b4f2f30c68b44cf8ae59b116d98f6c7) in parent group-v274074. [ 850.915319] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Creating folder: Instances. Parent ref: group-v274159. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 850.915568] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e2cbc91-bf8e-4237-b33e-cd52dd2f43c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.925615] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Created folder: Instances in parent group-v274159. [ 850.925895] env[61824]: DEBUG oslo.service.loopingcall [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.926118] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.926862] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-813db6b7-48dd-4453-8d5d-499b039930ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.957208] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.957208] env[61824]: value = "task-1275608" [ 850.957208] env[61824]: _type = "Task" [ 850.957208] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.974978] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275608, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.043026] env[61824]: ERROR nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [req-a7650ea3-09ce-40e8-a8a8-d209f8a1f578] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8d9ece0e-729c-4d69-b741-d89969200860. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a7650ea3-09ce-40e8-a8a8-d209f8a1f578"}]} [ 851.070360] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing inventories for resource provider 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 851.089564] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating ProviderTree inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 851.090783] env[61824]: DEBUG nova.compute.provider_tree [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 851.110717] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing aggregate associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, aggregates: None {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 851.119178] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52348a35-564e-a4ca-91a1-530481f01e3a, 'name': SearchDatastore_Task, 'duration_secs': 0.025824} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.119357] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.119530] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 31edad40-a743-4b65-8ef0-615a9ba2a225/31edad40-a743-4b65-8ef0-615a9ba2a225.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.120352] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ab0ed0a9-0826-4b4a-9749-67208432a08b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.128551] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 851.128551] env[61824]: value = "task-1275609" [ 851.128551] env[61824]: _type = "Task" [ 851.128551] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.135847] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing trait associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 851.142025] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.161198] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275605, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.232984] env[61824]: DEBUG nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.453677] env[61824]: DEBUG nova.network.neutron [-] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.474314] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275608, 'name': CreateVM_Task, 'duration_secs': 0.370559} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.474314] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.474314] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.474314] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.474641] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.475395] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52304d6e-2cd5-4005-93f6-5ccacdf482af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.485582] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 851.485582] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]528b9774-121b-ac71-cc08-284c0dca8782" [ 851.485582] env[61824]: _type = "Task" [ 851.485582] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.497440] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528b9774-121b-ac71-cc08-284c0dca8782, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.561934] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc4e690-1453-46a1-b192-ee7630123d9a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.573089] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2241a914-6195-409b-9167-ddf36ff8b207 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.610253] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f6602f-8e71-47f1-a4d6-5105a9c9ecd8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.618766] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f755e975-754f-433a-8aac-8e8f842b26b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.637101] env[61824]: DEBUG nova.compute.provider_tree [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 851.668356] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275609, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.675422] env[61824]: DEBUG oslo_vmware.api [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275605, 'name': PowerOnVM_Task, 'duration_secs': 0.836547} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.675422] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.675739] env[61824]: INFO nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Took 8.69 seconds to spawn the instance on the hypervisor. [ 851.675950] env[61824]: DEBUG nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.677342] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e079e9-bbe1-4952-bf54-60f3ecce7984 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.762435] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.847772] env[61824]: DEBUG nova.network.neutron [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Updated VIF entry in instance network info cache for port beb8cba8-71b3-40a5-ba5d-665a9770ec4d. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 851.848217] env[61824]: DEBUG nova.network.neutron [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Updating instance_info_cache with network_info: [{"id": "beb8cba8-71b3-40a5-ba5d-665a9770ec4d", "address": "fa:16:3e:75:25:03", "network": {"id": "9f8b7344-1620-45fe-ac3c-2d195c378f3e", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-883178807-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b4f2f30c68b44cf8ae59b116d98f6c7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49a4d142-3f97-47fe-b074-58923c46815e", "external-id": "nsx-vlan-transportzone-565", "segmentation_id": 565, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbeb8cba8-71", "ovs_interfaceid": "beb8cba8-71b3-40a5-ba5d-665a9770ec4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.960648] env[61824]: INFO nova.compute.manager [-] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Took 2.37 seconds to deallocate network for instance. [ 852.000477] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528b9774-121b-ac71-cc08-284c0dca8782, 'name': SearchDatastore_Task, 'duration_secs': 0.055345} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.000477] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.000938] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.001230] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.002354] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.002628] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.002838] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e53c7d26-fc0d-44df-b46e-549e218fa72c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.013416] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.013416] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.013416] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-010ead3f-e4e0-44d7-bce4-dc93e2dacc76 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.018885] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 852.018885] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1072c-7a94-904d-92f1-6e14e8f0932e" [ 852.018885] env[61824]: _type = "Task" [ 852.018885] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.027997] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1072c-7a94-904d-92f1-6e14e8f0932e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.146948] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556056} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.147213] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 31edad40-a743-4b65-8ef0-615a9ba2a225/31edad40-a743-4b65-8ef0-615a9ba2a225.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.147444] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.147702] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f8bff023-8b85-45eb-beee-4086c00d603f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.156863] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 852.156863] env[61824]: value = "task-1275610" [ 852.156863] env[61824]: _type = "Task" [ 852.156863] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.165895] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.198474] env[61824]: INFO nova.compute.manager [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Took 21.79 seconds to build instance. [ 852.201061] env[61824]: DEBUG nova.scheduler.client.report [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updated inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 852.201361] env[61824]: DEBUG nova.compute.provider_tree [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating resource provider 8d9ece0e-729c-4d69-b741-d89969200860 generation from 88 to 89 during operation: update_inventory {{(pid=61824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 852.201636] env[61824]: DEBUG nova.compute.provider_tree [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 852.257965] env[61824]: DEBUG nova.network.neutron [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Successfully updated port: bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.351341] env[61824]: DEBUG oslo_concurrency.lockutils [req-06336884-71ec-4b5a-8485-9d1b57133c00 req-9ee3b602-7ac0-49ea-9faa-5adbdfe68568 service nova] Releasing lock "refresh_cache-37c405be-62a1-4feb-8cae-ea1c924f4f40" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.498663] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "d9bb3172-c058-4e59-a719-d8aa340fba71" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.498958] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.519079] env[61824]: INFO nova.compute.manager [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Took 0.56 seconds to detach 1 volumes for instance. [ 852.524661] env[61824]: DEBUG nova.compute.manager [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Deleting volume: 32104921-3bcd-4691-86e7-9c3c0a08ed72 {{(pid=61824) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 852.533426] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e1072c-7a94-904d-92f1-6e14e8f0932e, 'name': SearchDatastore_Task, 'duration_secs': 0.008576} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.534267] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61205a4e-1665-49c7-8a64-926bbc10191b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.541124] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 852.541124] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]524582ed-d5ac-d9ab-1544-12117bde3f38" [ 852.541124] env[61824]: _type = "Task" [ 852.541124] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.551234] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524582ed-d5ac-d9ab-1544-12117bde3f38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.569229] env[61824]: DEBUG nova.compute.manager [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Received event network-vif-deleted-64000f4d-cb2c-4d63-95b7-97218e98723c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.569389] env[61824]: DEBUG nova.compute.manager [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-vif-plugged-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.569598] env[61824]: DEBUG oslo_concurrency.lockutils [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] Acquiring lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.569803] env[61824]: DEBUG oslo_concurrency.lockutils [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.570279] env[61824]: DEBUG oslo_concurrency.lockutils [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.570740] env[61824]: DEBUG nova.compute.manager [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] No waiting events found dispatching network-vif-plugged-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.570782] env[61824]: WARNING nova.compute.manager [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received unexpected event network-vif-plugged-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 for instance with vm_state building and task_state spawning. [ 852.570920] env[61824]: DEBUG nova.compute.manager [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.571124] env[61824]: DEBUG nova.compute.manager [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing instance network info cache due to event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 852.571325] env[61824]: DEBUG oslo_concurrency.lockutils [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.571463] env[61824]: DEBUG oslo_concurrency.lockutils [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.572032] env[61824]: DEBUG nova.network.neutron [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing network info cache for port bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 852.651868] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] Acquiring lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.651968] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] Acquired lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.652193] env[61824]: DEBUG nova.network.neutron [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.668294] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102385} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.668529] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.669365] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0de152-54e2-4504-8de6-d87795fdc694 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.692855] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 31edad40-a743-4b65-8ef0-615a9ba2a225/31edad40-a743-4b65-8ef0-615a9ba2a225.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.693500] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76d6fd76-8b48-4416-bf01-419ba03fd1dc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.708593] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf169336-b6bf-4f51-8f53-062ef046347a tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.306s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.709270] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.587s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.709751] env[61824]: DEBUG nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.712973] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.816s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.714432] env[61824]: INFO nova.compute.claims [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.721663] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 852.721663] env[61824]: value = "task-1275612" [ 852.721663] env[61824]: _type = "Task" [ 852.721663] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.730619] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.760632] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.005021] env[61824]: DEBUG nova.compute.utils [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.054023] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524582ed-d5ac-d9ab-1544-12117bde3f38, 'name': SearchDatastore_Task, 'duration_secs': 0.034121} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.054023] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.054023] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 37c405be-62a1-4feb-8cae-ea1c924f4f40/37c405be-62a1-4feb-8cae-ea1c924f4f40.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.054023] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2b0dbbc-34e4-42c7-b4e1-f38dce07025e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.059189] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 853.059189] env[61824]: value = "task-1275613" [ 853.059189] env[61824]: _type = "Task" [ 853.059189] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.067782] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.068937] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.154235] env[61824]: DEBUG nova.network.neutron [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.164435] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "b1a770f9-71ea-4f80-a562-c08a313b5753" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.164925] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.220092] env[61824]: DEBUG nova.compute.utils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.227338] env[61824]: DEBUG nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 853.227644] env[61824]: DEBUG nova.network.neutron [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 853.243248] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275612, 'name': ReconfigVM_Task, 'duration_secs': 0.289486} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.243757] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 31edad40-a743-4b65-8ef0-615a9ba2a225/31edad40-a743-4b65-8ef0-615a9ba2a225.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.244743] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ff2c008-4cfe-47a1-9e6f-9e2234a4e313 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.253272] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 853.253272] env[61824]: value = "task-1275614" [ 853.253272] env[61824]: _type = "Task" [ 853.253272] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.263309] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275614, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.275810] env[61824]: DEBUG nova.policy [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1d7f006ab244da48e1bfc38a5859932', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e23c7b193932478096b2fbf49a59fd7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 853.371059] env[61824]: DEBUG nova.network.neutron [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.509650] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.571058] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275613, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.572883] env[61824]: DEBUG nova.network.neutron [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Successfully created port: f28bdbaf-7121-4f64-916f-d52245e55d25 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.605888] env[61824]: DEBUG nova.network.neutron [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Updating instance_info_cache with network_info: [{"id": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "address": "fa:16:3e:a3:4c:22", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0fb98c3-23", "ovs_interfaceid": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.667578] env[61824]: DEBUG nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.727918] env[61824]: DEBUG nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.769427] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275614, 'name': Rename_Task, 'duration_secs': 0.200919} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.769610] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 853.769866] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2043d257-3dab-41ed-8f88-65912d449630 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.781052] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 853.781052] env[61824]: value = "task-1275615" [ 853.781052] env[61824]: _type = "Task" [ 853.781052] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.791493] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275615, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.873439] env[61824]: DEBUG oslo_concurrency.lockutils [req-97b5dbf4-9f02-4607-84c3-e4f6294ff5f3 req-e91c3ae6-71bc-4be6-bca5-1bf02c085790 service nova] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.873815] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.873964] env[61824]: DEBUG nova.network.neutron [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.041824] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4930a0bc-b6ea-43c8-8a87-da7f58fbf78b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.050273] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7ec9b2-ff27-4f65-974f-1a929d40908b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.087153] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e10a1b2-107d-4b35-a73e-1de5e6bd42fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.100174] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275613, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564881} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.100301] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 37c405be-62a1-4feb-8cae-ea1c924f4f40/37c405be-62a1-4feb-8cae-ea1c924f4f40.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.103072] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.103072] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072457af-dfc8-4b8b-8206-42fc3ff918a5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.110017] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7fbf1ef-2c60-4db4-a866-f1a6dddf1926 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.110017] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] Releasing lock "refresh_cache-01fd9faa-a0de-45b2-89bf-094c6d1ac155" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.110017] env[61824]: DEBUG nova.compute.manager [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Inject network info {{(pid=61824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 854.110252] env[61824]: DEBUG nova.compute.manager [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] network_info to inject: |[{"id": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "address": "fa:16:3e:a3:4c:22", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0fb98c3-23", "ovs_interfaceid": "e0fb98c3-2363-41c7-b63f-f3550b0050be", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 854.117232] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Reconfiguring VM instance to set the machine id {{(pid=61824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 854.118766] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd18c7fe-dfa8-4b26-a22d-2215dbf149ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.143846] env[61824]: DEBUG nova.compute.provider_tree [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.149831] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 854.149831] env[61824]: value = "task-1275616" [ 854.149831] env[61824]: _type = "Task" [ 854.149831] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.156837] env[61824]: DEBUG oslo_vmware.api [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] Waiting for the task: (returnval){ [ 854.156837] env[61824]: value = "task-1275617" [ 854.156837] env[61824]: _type = "Task" [ 854.156837] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.165651] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275616, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.180019] env[61824]: DEBUG oslo_vmware.api [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] Task: {'id': task-1275617, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.192446] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.293224] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275615, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.407671] env[61824]: DEBUG nova.network.neutron [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 854.559683] env[61824]: DEBUG nova.network.neutron [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.595364] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "d9bb3172-c058-4e59-a719-d8aa340fba71" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.596065] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.596065] env[61824]: INFO nova.compute.manager [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Attaching volume 6c0ac2ed-db17-4169-b133-158c17007c69 to /dev/sdb [ 854.633286] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39ec800-6358-490c-a24b-99dc71420de1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.640984] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b4566b-7280-4aca-a7f7-419eb6d452e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.647555] env[61824]: DEBUG nova.scheduler.client.report [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.656570] env[61824]: DEBUG nova.virt.block_device [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Updating existing volume attachment record: 6ef6c7f6-0ab4-43b5-9ed1-d7d41d8e6e7c {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 854.674118] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275616, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072857} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.680703] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.680703] env[61824]: DEBUG oslo_vmware.api [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] Task: {'id': task-1275617, 'name': ReconfigVM_Task, 'duration_secs': 0.165581} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.680703] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e973e8f-20e1-4867-b0ca-c03f52bd02ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.683563] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a18ac59c-cb5e-4383-ab88-06d4f90e169a tempest-ServersAdminTestJSON-1608206215 tempest-ServersAdminTestJSON-1608206215-project-admin] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Reconfigured VM instance to set the machine id {{(pid=61824) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 854.707192] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 37c405be-62a1-4feb-8cae-ea1c924f4f40/37c405be-62a1-4feb-8cae-ea1c924f4f40.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.707554] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6e200d3-1cfa-45c2-9589-68dd379e7e6c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.728469] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 854.728469] env[61824]: value = "task-1275618" [ 854.728469] env[61824]: _type = "Task" [ 854.728469] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.735594] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275618, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.740915] env[61824]: DEBUG nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.785021] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='33907b9f6abc18985ba3d060dc06462f',container_format='bare',created_at=2024-10-12T10:58:02Z,direct_url=,disk_format='vmdk',id=619a0c61-29bb-40c8-bc1c-82d0621b6650,min_disk=1,min_ram=0,name='tempest-test-snap-1532785173',owner='e23c7b193932478096b2fbf49a59fd7e',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-12T10:58:16Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.785021] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.785021] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.785021] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.785295] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.785295] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.785295] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.785295] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.785295] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.785435] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.785435] env[61824]: DEBUG nova.virt.hardware [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.785889] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ef53b5-7f5d-4a76-9901-f513d9bbb16d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.801562] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8d7f09-41a7-473c-bb0d-1ab3f02f5e17 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.803825] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275615, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.062762] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.063050] env[61824]: DEBUG nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Instance network_info: |[{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 855.063522] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:95:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6edb8eae-1113-49d0-84f7-9fd9f82b26fb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd35af69-df49-4c0a-a798-57bd9c8f6fe4', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.071575] env[61824]: DEBUG oslo.service.loopingcall [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.072753] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.073856] env[61824]: DEBUG nova.compute.manager [req-c1560d8a-17c4-421a-bb61-26ba9a704b25 req-452b5b90-442a-413d-87cf-5a52815a49ef service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Received event network-vif-plugged-f28bdbaf-7121-4f64-916f-d52245e55d25 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.074073] env[61824]: DEBUG oslo_concurrency.lockutils [req-c1560d8a-17c4-421a-bb61-26ba9a704b25 req-452b5b90-442a-413d-87cf-5a52815a49ef service nova] Acquiring lock "5806faad-2bab-4fda-a131-566af50c4d7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.074285] env[61824]: DEBUG oslo_concurrency.lockutils [req-c1560d8a-17c4-421a-bb61-26ba9a704b25 req-452b5b90-442a-413d-87cf-5a52815a49ef service nova] Lock "5806faad-2bab-4fda-a131-566af50c4d7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.074452] env[61824]: DEBUG oslo_concurrency.lockutils [req-c1560d8a-17c4-421a-bb61-26ba9a704b25 req-452b5b90-442a-413d-87cf-5a52815a49ef service nova] Lock "5806faad-2bab-4fda-a131-566af50c4d7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.074619] env[61824]: DEBUG nova.compute.manager [req-c1560d8a-17c4-421a-bb61-26ba9a704b25 req-452b5b90-442a-413d-87cf-5a52815a49ef service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] No waiting events found dispatching network-vif-plugged-f28bdbaf-7121-4f64-916f-d52245e55d25 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 855.074784] env[61824]: WARNING nova.compute.manager [req-c1560d8a-17c4-421a-bb61-26ba9a704b25 req-452b5b90-442a-413d-87cf-5a52815a49ef service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Received unexpected event network-vif-plugged-f28bdbaf-7121-4f64-916f-d52245e55d25 for instance with vm_state building and task_state spawning. [ 855.075239] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87c23647-870a-425a-9a41-e4c954f75b50 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.094855] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.094855] env[61824]: value = "task-1275622" [ 855.094855] env[61824]: _type = "Task" [ 855.094855] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.103896] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275622, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.127032] env[61824]: DEBUG nova.network.neutron [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Successfully updated port: f28bdbaf-7121-4f64-916f-d52245e55d25 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 855.152722] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.153357] env[61824]: DEBUG nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 855.156088] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.959s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.237336] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275618, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.296043] env[61824]: DEBUG oslo_vmware.api [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275615, 'name': PowerOnVM_Task, 'duration_secs': 1.075781} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.296043] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.296043] env[61824]: INFO nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Took 9.92 seconds to spawn the instance on the hypervisor. [ 855.296270] env[61824]: DEBUG nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.296944] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869507ba-c264-4293-b58b-ea6fcd9da581 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.604800] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275622, 'name': CreateVM_Task, 'duration_secs': 0.451777} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.605161] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.605658] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.605828] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.606189] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.606442] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8af2e3e6-3626-4631-993b-2cf199cd23ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.611097] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 855.611097] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525d908f-651e-1aa5-6a77-4aa20d9c35ef" [ 855.611097] env[61824]: _type = "Task" [ 855.611097] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.618381] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525d908f-651e-1aa5-6a77-4aa20d9c35ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.632953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "refresh_cache-5806faad-2bab-4fda-a131-566af50c4d7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.633104] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "refresh_cache-5806faad-2bab-4fda-a131-566af50c4d7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.633268] env[61824]: DEBUG nova.network.neutron [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 855.658968] env[61824]: DEBUG nova.compute.utils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 855.660235] env[61824]: DEBUG nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 855.660418] env[61824]: DEBUG nova.network.neutron [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 855.707141] env[61824]: DEBUG nova.policy [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1eb545ac2cf4f0d9ccddb040b17372b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e756c6817698464a9da391d4dd364946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 855.739234] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275618, 'name': ReconfigVM_Task, 'duration_secs': 0.687682} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.739234] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 37c405be-62a1-4feb-8cae-ea1c924f4f40/37c405be-62a1-4feb-8cae-ea1c924f4f40.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.739479] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b18970d6-4f4c-4774-b0e1-49621c10307b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.745558] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 855.745558] env[61824]: value = "task-1275623" [ 855.745558] env[61824]: _type = "Task" [ 855.745558] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.754672] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275623, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.817282] env[61824]: INFO nova.compute.manager [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Took 20.88 seconds to build instance. [ 855.896808] env[61824]: INFO nova.compute.manager [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Rebuilding instance [ 855.943150] env[61824]: DEBUG nova.compute.manager [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.944072] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34cdf7ee-033b-48fa-8504-2036346cb958 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.986814] env[61824]: DEBUG nova.network.neutron [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Successfully created port: cca34907-2b24-40e6-8cab-24bacb75db19 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.121756] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525d908f-651e-1aa5-6a77-4aa20d9c35ef, 'name': SearchDatastore_Task, 'duration_secs': 0.009863} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.122088] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.122325] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.122563] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.122715] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.122912] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.123524] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62bdde15-492b-4a7d-8eb6-0f7f68758eda {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.131886] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.131978] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 856.132833] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1fd3dea-aad8-461e-bd5c-d8c42f27c772 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.140854] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 856.140854] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5201d2e2-1ede-542e-32c9-f17e368ac18f" [ 856.140854] env[61824]: _type = "Task" [ 856.140854] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.149867] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5201d2e2-1ede-542e-32c9-f17e368ac18f, 'name': SearchDatastore_Task, 'duration_secs': 0.008721} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.150305] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5ac906c-dac9-4988-944d-a91e3c40044e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.155795] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 856.155795] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]523da3c9-34a3-d0b5-3f8d-2a0b36f33707" [ 856.155795] env[61824]: _type = "Task" [ 856.155795] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.165119] env[61824]: DEBUG nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 856.167775] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523da3c9-34a3-d0b5-3f8d-2a0b36f33707, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.171406] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Applying migration context for instance a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 as it has an incoming, in-progress migration de82b366-9a8e-489e-bd6f-2729751101dc. Migration status is confirming {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 856.173224] env[61824]: INFO nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating resource usage from migration de82b366-9a8e-489e-bd6f-2729751101dc [ 856.185581] env[61824]: DEBUG nova.network.neutron [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 856.207178] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7e73f8b5-a138-4455-a392-9a2b2b860558 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207178] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7f6a4446-86e1-44c7-ab42-297b033dbace actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207178] env[61824]: WARNING nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 6235a9b9-19c0-4917-938a-bca82a2ca05e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 856.207178] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance d9bb3172-c058-4e59-a719-d8aa340fba71 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207440] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7be79422-c624-4b4a-884f-29b13b9d81d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207440] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance cbded637-56b9-4048-b66e-1edf6311f435 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207440] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 2495fcb4-28cb-49c3-90fe-c84072466287 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207440] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance d6576f45-35f6-48dd-9160-80fc53d8b6b2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207563] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 5569eb0a-fd42-4e2d-b805-bb40e3264a65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207563] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Migration de82b366-9a8e-489e-bd6f-2729751101dc is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 856.207563] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207563] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207691] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 21759f11-80e9-419d-9f65-ca4767d1593b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207691] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 01fd9faa-a0de-45b2-89bf-094c6d1ac155 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207691] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 31edad40-a743-4b65-8ef0-615a9ba2a225 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207691] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 37c405be-62a1-4feb-8cae-ea1c924f4f40 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207808] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance aa9c9eb2-b411-4eca-9f8d-3df377ffe351 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207861] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 5806faad-2bab-4fda-a131-566af50c4d7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.207993] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance cabad7a6-206f-4e02-be3c-e18654396b7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 856.261113] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275623, 'name': Rename_Task, 'duration_secs': 0.15825} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.261407] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.261727] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f98d524-6115-4756-923b-b8f5ae9af35d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.268550] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 856.268550] env[61824]: value = "task-1275624" [ 856.268550] env[61824]: _type = "Task" [ 856.268550] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.276592] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.320217] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e5580ceb-7413-4e52-99aa-2f4d6d87f632 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.388s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.399184] env[61824]: DEBUG nova.network.neutron [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Updating instance_info_cache with network_info: [{"id": "f28bdbaf-7121-4f64-916f-d52245e55d25", "address": "fa:16:3e:6f:32:ab", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf28bdbaf-71", "ovs_interfaceid": "f28bdbaf-7121-4f64-916f-d52245e55d25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.455526] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 856.455907] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c4dbe6e-9531-4f9a-b652-c8866d095397 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.463933] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 856.463933] env[61824]: value = "task-1275625" [ 856.463933] env[61824]: _type = "Task" [ 856.463933] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.473564] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.667299] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523da3c9-34a3-d0b5-3f8d-2a0b36f33707, 'name': SearchDatastore_Task, 'duration_secs': 0.0084} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.667652] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.667897] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] aa9c9eb2-b411-4eca-9f8d-3df377ffe351/aa9c9eb2-b411-4eca-9f8d-3df377ffe351.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.668164] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee240b1f-112c-4137-bfb4-17b97ccd00fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.677751] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 856.677751] env[61824]: value = "task-1275626" [ 856.677751] env[61824]: _type = "Task" [ 856.677751] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.686607] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275626, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.713379] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance c6552176-1128-473b-ac4a-f3b55428fd9e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 856.780264] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275624, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.901560] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "refresh_cache-5806faad-2bab-4fda-a131-566af50c4d7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.902042] env[61824]: DEBUG nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Instance network_info: |[{"id": "f28bdbaf-7121-4f64-916f-d52245e55d25", "address": "fa:16:3e:6f:32:ab", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf28bdbaf-71", "ovs_interfaceid": "f28bdbaf-7121-4f64-916f-d52245e55d25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 856.902650] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:32:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7f41333-42ee-47f3-936c-d6701ab786d2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f28bdbaf-7121-4f64-916f-d52245e55d25', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.911203] env[61824]: DEBUG oslo.service.loopingcall [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.911526] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.911686] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb685910-b5c5-4bd4-8c55-b030e328cdab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.932422] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.932422] env[61824]: value = "task-1275627" [ 856.932422] env[61824]: _type = "Task" [ 856.932422] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.936319] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.936567] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.936760] env[61824]: INFO nova.compute.manager [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Shelving [ 856.945033] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275627, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.977215] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275625, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.103420] env[61824]: DEBUG nova.compute.manager [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Received event network-changed-f28bdbaf-7121-4f64-916f-d52245e55d25 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.103618] env[61824]: DEBUG nova.compute.manager [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Refreshing instance network info cache due to event network-changed-f28bdbaf-7121-4f64-916f-d52245e55d25. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 857.103837] env[61824]: DEBUG oslo_concurrency.lockutils [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] Acquiring lock "refresh_cache-5806faad-2bab-4fda-a131-566af50c4d7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.104010] env[61824]: DEBUG oslo_concurrency.lockutils [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] Acquired lock "refresh_cache-5806faad-2bab-4fda-a131-566af50c4d7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.104357] env[61824]: DEBUG nova.network.neutron [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Refreshing network info cache for port f28bdbaf-7121-4f64-916f-d52245e55d25 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 857.177236] env[61824]: DEBUG nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 857.191572] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275626, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.205320] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 857.205686] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 857.205898] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 857.206129] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 857.206280] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 857.206427] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 857.206652] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 857.206803] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 857.207669] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 857.207669] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 857.207669] env[61824]: DEBUG nova.virt.hardware [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 857.209030] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a125697b-6ed5-4896-9e2c-052bd93cd06c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.216479] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance b1a770f9-71ea-4f80-a562-c08a313b5753 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 857.216756] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 857.216903] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4032MB phys_disk=200GB used_disk=18GB total_vcpus=48 used_vcpus=18 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 857.221173] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb283dd-8203-44ca-85ff-ca7d0ed439ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.279776] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275624, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.448765] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275627, 'name': CreateVM_Task, 'duration_secs': 0.405254} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.450915] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 857.451448] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.452065] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.452240] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.452624] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.452851] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b9542a0-b6a4-4a1e-8ae0-2426c4270e14 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.457262] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec25e948-4c2d-4e9b-be34-31cfab864839 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.461816] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "31edad40-a743-4b65-8ef0-615a9ba2a225" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.462064] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.462275] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "31edad40-a743-4b65-8ef0-615a9ba2a225-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.462459] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.462631] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.464213] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 857.464213] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52aa7ad7-5bb0-4b34-cbbd-a93a9aaf1f7c" [ 857.464213] env[61824]: _type = "Task" [ 857.464213] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.464641] env[61824]: INFO nova.compute.manager [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Terminating instance [ 857.467262] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 857.467262] env[61824]: value = "task-1275629" [ 857.467262] env[61824]: _type = "Task" [ 857.467262] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.467831] env[61824]: DEBUG nova.compute.manager [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 857.468229] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 857.476752] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8722fd4-b782-43e1-a58c-7127fea55334 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.487718] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275625, 'name': PowerOffVM_Task, 'duration_secs': 0.615428} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.495668] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.495957] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 857.496277] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275629, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.496668] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.500639] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e0fd4e-50e7-4d69-802e-d46d398e984b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.503385] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.503620] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Processing image 619a0c61-29bb-40c8-bc1c-82d0621b6650 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.503947] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650/619a0c61-29bb-40c8-bc1c-82d0621b6650.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.503984] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650/619a0c61-29bb-40c8-bc1c-82d0621b6650.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.504188] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.504406] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-29942b4d-7c64-45a9-b258-b291ad89c426 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.506649] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5709687a-bea6-4a3d-8786-f8efe24f89af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.516018] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 857.519339] env[61824]: DEBUG nova.network.neutron [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Successfully updated port: cca34907-2b24-40e6-8cab-24bacb75db19 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.520502] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c6a6f3ac-2850-4052-b65d-285e343ab4c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.522733] env[61824]: DEBUG oslo_vmware.api [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 857.522733] env[61824]: value = "task-1275630" [ 857.522733] env[61824]: _type = "Task" [ 857.522733] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.526450] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.526945] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 857.533074] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31c29ca6-9034-41ff-9a7c-7090529dbf45 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.539665] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 857.539665] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c0eea0-4670-e365-9050-9017312caeef" [ 857.539665] env[61824]: _type = "Task" [ 857.539665] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.543613] env[61824]: DEBUG oslo_vmware.api [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.548802] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c87679-e831-45fc-9a89-f50511bd63ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.560448] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1922fe77-3cb1-492a-9ec4-9f5b91218263 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.563980] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Preparing fetch location {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 857.564402] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Fetch image to [datastore2] OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4/OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4.vmdk {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 857.564502] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Downloading stream optimized image 619a0c61-29bb-40c8-bc1c-82d0621b6650 to [datastore2] OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4/OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4.vmdk on the data store datastore2 as vApp {{(pid=61824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 857.564635] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Downloading image file data 619a0c61-29bb-40c8-bc1c-82d0621b6650 to the ESX as VM named 'OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4' {{(pid=61824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 857.597254] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302538d9-7c04-4223-85c9-8be7a96bbf23 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.622709] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4c5ee8-ea4e-44c1-8cfd-d490969a117a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.630895] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 857.631151] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 857.631333] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleting the datastore file [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.633573] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-46182839-30d8-4ee7-94a4-825c3f63ea4b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.643316] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.650846] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 857.650846] env[61824]: value = "task-1275632" [ 857.650846] env[61824]: _type = "Task" [ 857.650846] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.660241] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.674538] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 857.674538] env[61824]: value = "resgroup-9" [ 857.674538] env[61824]: _type = "ResourcePool" [ 857.674538] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 857.674886] env[61824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-74b1f372-6015-42a4-8efb-1e7f7eba58be {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.701668] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275626, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.611249} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.705218] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] aa9c9eb2-b411-4eca-9f8d-3df377ffe351/aa9c9eb2-b411-4eca-9f8d-3df377ffe351.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 857.705455] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.705769] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lease: (returnval){ [ 857.705769] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52945e59-3aff-fcee-2b59-2745fa4ca803" [ 857.705769] env[61824]: _type = "HttpNfcLease" [ 857.705769] env[61824]: } obtained for vApp import into resource pool (val){ [ 857.705769] env[61824]: value = "resgroup-9" [ 857.705769] env[61824]: _type = "ResourcePool" [ 857.705769] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 857.706113] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the lease: (returnval){ [ 857.706113] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52945e59-3aff-fcee-2b59-2745fa4ca803" [ 857.706113] env[61824]: _type = "HttpNfcLease" [ 857.706113] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 857.706278] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-409bf178-9236-45a9-b176-034fd0d79455 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.716244] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 857.716244] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52945e59-3aff-fcee-2b59-2745fa4ca803" [ 857.716244] env[61824]: _type = "HttpNfcLease" [ 857.716244] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 857.717561] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 857.717561] env[61824]: value = "task-1275634" [ 857.717561] env[61824]: _type = "Task" [ 857.717561] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.728105] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275634, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.780090] env[61824]: DEBUG oslo_vmware.api [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275624, 'name': PowerOnVM_Task, 'duration_secs': 1.14915} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.780403] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.780641] env[61824]: INFO nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Took 10.04 seconds to spawn the instance on the hypervisor. [ 857.780830] env[61824]: DEBUG nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.781678] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236cb4cd-902e-4851-85f0-b2e2a343495b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.872346] env[61824]: DEBUG nova.network.neutron [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Updated VIF entry in instance network info cache for port f28bdbaf-7121-4f64-916f-d52245e55d25. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.872695] env[61824]: DEBUG nova.network.neutron [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Updating instance_info_cache with network_info: [{"id": "f28bdbaf-7121-4f64-916f-d52245e55d25", "address": "fa:16:3e:6f:32:ab", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf28bdbaf-71", "ovs_interfaceid": "f28bdbaf-7121-4f64-916f-d52245e55d25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.986061] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275629, 'name': PowerOffVM_Task, 'duration_secs': 0.310199} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.986061] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.986832] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66feb59a-1048-4627-82af-e52155a81dae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.005784] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08ba6e8-4299-474c-b289-ac58ae498b07 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.026399] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "refresh_cache-cabad7a6-206f-4e02-be3c-e18654396b7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.026399] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "refresh_cache-cabad7a6-206f-4e02-be3c-e18654396b7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.026575] env[61824]: DEBUG nova.network.neutron [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.037190] env[61824]: DEBUG oslo_vmware.api [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275630, 'name': PowerOffVM_Task, 'duration_secs': 0.291845} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.038150] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.038377] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 858.038639] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-11a2807e-5f56-41f1-823f-b6a3f088d9b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.129840] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 858.130080] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 858.130320] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Deleting the datastore file [datastore1] 31edad40-a743-4b65-8ef0-615a9ba2a225 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.130613] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb20c423-186a-4f06-b1ef-f6d5315c6d47 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.140804] env[61824]: DEBUG oslo_vmware.api [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for the task: (returnval){ [ 858.140804] env[61824]: value = "task-1275636" [ 858.140804] env[61824]: _type = "Task" [ 858.140804] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.145759] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.152171] env[61824]: DEBUG oslo_vmware.api [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275636, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.162466] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275632, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14491} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.163616] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.163877] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 858.164135] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 858.219340] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.219340] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52945e59-3aff-fcee-2b59-2745fa4ca803" [ 858.219340] env[61824]: _type = "HttpNfcLease" [ 858.219340] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 858.226230] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275634, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.302690] env[61824]: INFO nova.compute.manager [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Took 18.96 seconds to build instance. [ 858.376025] env[61824]: DEBUG oslo_concurrency.lockutils [req-9778409c-6fd2-4f14-bd08-58d48488b94d req-b451ec57-8398-41e7-8e62-4e7fee87fac5 service nova] Releasing lock "refresh_cache-5806faad-2bab-4fda-a131-566af50c4d7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.518964] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Creating Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 858.519328] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e5e28ced-701c-4dd0-bdfe-f48376499cde {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.526455] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 858.526455] env[61824]: value = "task-1275637" [ 858.526455] env[61824]: _type = "Task" [ 858.526455] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.536839] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275637, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.578863] env[61824]: DEBUG nova.network.neutron [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.651939] env[61824]: DEBUG oslo_vmware.api [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Task: {'id': task-1275636, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.477096} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.651939] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.652066] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 858.652247] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 858.652427] env[61824]: INFO nova.compute.manager [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Took 1.18 seconds to destroy the instance on the hypervisor. [ 858.652671] env[61824]: DEBUG oslo.service.loopingcall [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.654144] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 858.654238] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.498s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.654402] env[61824]: DEBUG nova.compute.manager [-] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 858.654498] env[61824]: DEBUG nova.network.neutron [-] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 858.656121] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.999s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.726600] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 858.726600] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52945e59-3aff-fcee-2b59-2745fa4ca803" [ 858.726600] env[61824]: _type = "HttpNfcLease" [ 858.726600] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 858.729402] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 858.729402] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52945e59-3aff-fcee-2b59-2745fa4ca803" [ 858.729402] env[61824]: _type = "HttpNfcLease" [ 858.729402] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 858.730130] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0132bb54-7be8-4a1b-85e4-283a4d963f11 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.738906] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275634, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.749094] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520bd40e-1fa6-ea90-d0df-daf1871af2c3/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 858.750025] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520bd40e-1fa6-ea90-d0df-daf1871af2c3/disk-0.vmdk. {{(pid=61824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 858.812824] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bdf502ea-14a3-48a8-b7f3-d38049bb18df tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.479s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.813980] env[61824]: DEBUG nova.network.neutron [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Updating instance_info_cache with network_info: [{"id": "cca34907-2b24-40e6-8cab-24bacb75db19", "address": "fa:16:3e:55:4a:21", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca34907-2b", "ovs_interfaceid": "cca34907-2b24-40e6-8cab-24bacb75db19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.823632] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8f0b85a1-b2ab-47a0-a319-4159bed117ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.036892] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275637, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.211089] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.211635] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.211635] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.211817] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.211921] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.212154] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.212403] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.212613] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.213826] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.213826] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.213826] env[61824]: DEBUG nova.virt.hardware [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.217755] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151185d7-7ad8-49ad-8f88-ad4e0b89972e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.223235] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 859.223235] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274163', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'name': 'volume-6c0ac2ed-db17-4169-b133-158c17007c69', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9bb3172-c058-4e59-a719-d8aa340fba71', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'serial': '6c0ac2ed-db17-4169-b133-158c17007c69'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 859.224228] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f36c43-9b7c-4619-ad1b-4d1e5f442435 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.256177] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e81940-f854-4101-b019-1f89779baa78 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.270891] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9e125f3-1e72-4da7-a553-edf37c9e9d87 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.273703] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275634, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.291482] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:52:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6076d24d-3c8e-4bbb-ba96-a08fb27a73cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51fba888-670e-4f85-97a5-0133e0624342', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.299130] env[61824]: DEBUG oslo.service.loopingcall [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.323965] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.333042] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] volume-6c0ac2ed-db17-4169-b133-158c17007c69/volume-6c0ac2ed-db17-4169-b133-158c17007c69.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.333605] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "refresh_cache-cabad7a6-206f-4e02-be3c-e18654396b7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.333927] env[61824]: DEBUG nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Instance network_info: |[{"id": "cca34907-2b24-40e6-8cab-24bacb75db19", "address": "fa:16:3e:55:4a:21", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca34907-2b", "ovs_interfaceid": "cca34907-2b24-40e6-8cab-24bacb75db19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 859.338324] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa129699-3e03-49a8-979f-d199b21f91b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.354970] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b6943bc-518e-4c1e-8350-2f5969b04d04 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.368319] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:4a:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cca34907-2b24-40e6-8cab-24bacb75db19', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.375825] env[61824]: DEBUG oslo.service.loopingcall [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.376911] env[61824]: DEBUG nova.network.neutron [-] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.380442] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.382291] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-864e3756-71dc-43fc-afe8-76a7371ec990 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.408200] env[61824]: DEBUG oslo_vmware.api [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 859.408200] env[61824]: value = "task-1275639" [ 859.408200] env[61824]: _type = "Task" [ 859.408200] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.408812] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.408812] env[61824]: value = "task-1275638" [ 859.408812] env[61824]: _type = "Task" [ 859.408812] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.419751] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.419751] env[61824]: value = "task-1275640" [ 859.419751] env[61824]: _type = "Task" [ 859.419751] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.431144] env[61824]: DEBUG oslo_vmware.api [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275639, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.438726] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275638, 'name': CreateVM_Task} progress is 15%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.446023] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275640, 'name': CreateVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.516162] env[61824]: DEBUG nova.compute.manager [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Received event network-vif-plugged-cca34907-2b24-40e6-8cab-24bacb75db19 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.516162] env[61824]: DEBUG oslo_concurrency.lockutils [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] Acquiring lock "cabad7a6-206f-4e02-be3c-e18654396b7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.516345] env[61824]: DEBUG oslo_concurrency.lockutils [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.516446] env[61824]: DEBUG oslo_concurrency.lockutils [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.516765] env[61824]: DEBUG nova.compute.manager [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] No waiting events found dispatching network-vif-plugged-cca34907-2b24-40e6-8cab-24bacb75db19 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.516967] env[61824]: WARNING nova.compute.manager [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Received unexpected event network-vif-plugged-cca34907-2b24-40e6-8cab-24bacb75db19 for instance with vm_state building and task_state spawning. [ 859.516967] env[61824]: DEBUG nova.compute.manager [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Received event network-changed-cca34907-2b24-40e6-8cab-24bacb75db19 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.517324] env[61824]: DEBUG nova.compute.manager [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Refreshing instance network info cache due to event network-changed-cca34907-2b24-40e6-8cab-24bacb75db19. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 859.517498] env[61824]: DEBUG oslo_concurrency.lockutils [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] Acquiring lock "refresh_cache-cabad7a6-206f-4e02-be3c-e18654396b7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.517596] env[61824]: DEBUG oslo_concurrency.lockutils [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] Acquired lock "refresh_cache-cabad7a6-206f-4e02-be3c-e18654396b7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.517756] env[61824]: DEBUG nova.network.neutron [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Refreshing network info cache for port cca34907-2b24-40e6-8cab-24bacb75db19 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 859.540670] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275637, 'name': CreateSnapshot_Task, 'duration_secs': 0.909932} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.541350] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Created Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 859.541817] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a56f62-2c7f-4849-9c85-40e12ad011a2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.600762] env[61824]: DEBUG nova.compute.manager [req-4e7c133c-47a6-4b5d-9807-e8a59a0f55af req-761fa46a-0bee-46c4-baba-7af2a45c57e5 service nova] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Received event network-vif-deleted-8ff3d743-0e3d-47f7-85ad-8981881e11ac {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.651786] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Completed reading data from the image iterator. {{(pid=61824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 859.652096] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520bd40e-1fa6-ea90-d0df-daf1871af2c3/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 859.653658] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef8950c-ceb7-4e4c-8446-889ceae3782b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.665091] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520bd40e-1fa6-ea90-d0df-daf1871af2c3/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 859.665284] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520bd40e-1fa6-ea90-d0df-daf1871af2c3/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 859.665530] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-55f2d28b-208b-4391-8279-256091dea1e9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.690563] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfeb227-040a-460f-b43c-337936073f03 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.702093] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228e90ad-203d-445f-95d9-4d4fb0a475c5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.753139] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1bb5b2-a1f7-48e0-beb7-0a2e4733aaf1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.764896] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275634, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.916135} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.767956] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.769242] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d66c6cb-4c64-4ac8-bb19-758291ed1316 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.773699] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2ad596-38c7-4b97-add4-f800b67ca1eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.796408] env[61824]: DEBUG nova.compute.provider_tree [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.828017] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] aa9c9eb2-b411-4eca-9f8d-3df377ffe351/aa9c9eb2-b411-4eca-9f8d-3df377ffe351.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.828017] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e995cc7-3ad3-4c43-8586-518bcd40ee2d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.847124] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 859.847124] env[61824]: value = "task-1275641" [ 859.847124] env[61824]: _type = "Task" [ 859.847124] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.862186] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275641, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.880637] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520bd40e-1fa6-ea90-d0df-daf1871af2c3/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 859.880870] env[61824]: INFO nova.virt.vmwareapi.images [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Downloaded image file data 619a0c61-29bb-40c8-bc1c-82d0621b6650 [ 859.881723] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea4b3e1-ea97-43c3-bd39-0329dc2e4e26 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.884604] env[61824]: INFO nova.compute.manager [-] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Took 1.23 seconds to deallocate network for instance. [ 859.901758] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-028bb96c-f35b-423e-b54d-8d949d65bcc1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.923890] env[61824]: DEBUG oslo_vmware.api [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275639, 'name': ReconfigVM_Task, 'duration_secs': 0.485732} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.930907] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Reconfigured VM instance instance-0000003f to attach disk [datastore2] volume-6c0ac2ed-db17-4169-b133-158c17007c69/volume-6c0ac2ed-db17-4169-b133-158c17007c69.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.936375] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275638, 'name': CreateVM_Task, 'duration_secs': 0.354906} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.937280] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e9da719-87cd-478b-a7aa-7c3f8229a5b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.947952] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.950618] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.950799] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.951202] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.951513] env[61824]: INFO nova.virt.vmwareapi.images [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] The imported VM was unregistered [ 859.953984] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Caching image {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 859.954229] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating directory with path [datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.958019] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7489bf19-fbe1-4087-88b8-3eaaeb1d182c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.960185] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f58b00db-e104-4179-b21c-e870346dd2bd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.962792] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275640, 'name': CreateVM_Task, 'duration_secs': 0.364896} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.964070] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.964429] env[61824]: DEBUG oslo_vmware.api [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 859.964429] env[61824]: value = "task-1275643" [ 859.964429] env[61824]: _type = "Task" [ 859.964429] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.968925] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.971011] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 859.971011] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52637cf6-263f-d250-ecc1-1443e460df0f" [ 859.971011] env[61824]: _type = "Task" [ 859.971011] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.979177] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Created directory with path [datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.979594] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4/OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4.vmdk to [datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650/619a0c61-29bb-40c8-bc1c-82d0621b6650.vmdk. {{(pid=61824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 859.979737] env[61824]: DEBUG oslo_vmware.api [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275643, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.980293] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-107d81eb-f92d-41df-a878-b5c99412c78c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.986075] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52637cf6-263f-d250-ecc1-1443e460df0f, 'name': SearchDatastore_Task, 'duration_secs': 0.00963} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.986718] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.986993] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.987241] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.987410] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.987608] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.987936] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.988275] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.988508] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-63d3c339-e08f-4d04-940c-718875b9a205 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.991427] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aec849a5-e604-4678-807f-4a84d9b4ba7e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.993155] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 859.993155] env[61824]: value = "task-1275644" [ 859.993155] env[61824]: _type = "Task" [ 859.993155] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.999410] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 859.999410] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5251f754-fbc6-37c3-4ab5-b33ef788a69c" [ 859.999410] env[61824]: _type = "Task" [ 859.999410] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.001387] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.001661] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.008523] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34945afa-1252-4a17-8503-8f88908f8392 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.011158] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275644, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.016954] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5251f754-fbc6-37c3-4ab5-b33ef788a69c, 'name': SearchDatastore_Task, 'duration_secs': 0.009637} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.018390] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.018678] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.018963] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.019354] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 860.019354] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5228d988-a5d4-d9ed-becc-ea20c4e211bc" [ 860.019354] env[61824]: _type = "Task" [ 860.019354] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.030787] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5228d988-a5d4-d9ed-becc-ea20c4e211bc, 'name': SearchDatastore_Task, 'duration_secs': 0.008797} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.032414] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94c236da-c013-4727-adb4-24df6b4d852f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.040022] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 860.040022] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]522f4bf9-3da8-360b-14ce-3282aab43cee" [ 860.040022] env[61824]: _type = "Task" [ 860.040022] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.047026] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522f4bf9-3da8-360b-14ce-3282aab43cee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.070341] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Creating linked-clone VM from snapshot {{(pid=61824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 860.071092] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bb80b62b-68e2-4be9-887f-d92e77d1cf86 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.082040] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 860.082040] env[61824]: value = "task-1275645" [ 860.082040] env[61824]: _type = "Task" [ 860.082040] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.090676] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275645, 'name': CloneVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.258539] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "37c405be-62a1-4feb-8cae-ea1c924f4f40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.258841] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.259321] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "37c405be-62a1-4feb-8cae-ea1c924f4f40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.259562] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.259746] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.262494] env[61824]: INFO nova.compute.manager [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Terminating instance [ 860.266169] env[61824]: DEBUG nova.compute.manager [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 860.266418] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 860.267393] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f4f51c-d0b0-4547-a51a-29cfd991e57c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.276457] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.276797] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f950e9a-f8c1-4f57-bf9c-37fe40023fab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.283366] env[61824]: DEBUG oslo_vmware.api [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 860.283366] env[61824]: value = "task-1275646" [ 860.283366] env[61824]: _type = "Task" [ 860.283366] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.295610] env[61824]: DEBUG oslo_vmware.api [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275646, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.328535] env[61824]: DEBUG nova.scheduler.client.report [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.336029] env[61824]: DEBUG nova.network.neutron [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Updated VIF entry in instance network info cache for port cca34907-2b24-40e6-8cab-24bacb75db19. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 860.336029] env[61824]: DEBUG nova.network.neutron [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Updating instance_info_cache with network_info: [{"id": "cca34907-2b24-40e6-8cab-24bacb75db19", "address": "fa:16:3e:55:4a:21", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcca34907-2b", "ovs_interfaceid": "cca34907-2b24-40e6-8cab-24bacb75db19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.358936] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275641, 'name': ReconfigVM_Task, 'duration_secs': 0.335598} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.359279] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Reconfigured VM instance instance-0000004b to attach disk [datastore1] aa9c9eb2-b411-4eca-9f8d-3df377ffe351/aa9c9eb2-b411-4eca-9f8d-3df377ffe351.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.360252] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f24fe32-84d2-45fd-806d-6ff072502d97 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.367555] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 860.367555] env[61824]: value = "task-1275647" [ 860.367555] env[61824]: _type = "Task" [ 860.367555] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.378046] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275647, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.403165] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.478723] env[61824]: DEBUG oslo_vmware.api [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275643, 'name': ReconfigVM_Task, 'duration_secs': 0.15581} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.479133] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274163', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'name': 'volume-6c0ac2ed-db17-4169-b133-158c17007c69', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9bb3172-c058-4e59-a719-d8aa340fba71', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'serial': '6c0ac2ed-db17-4169-b133-158c17007c69'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 860.504619] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275644, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.550158] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522f4bf9-3da8-360b-14ce-3282aab43cee, 'name': SearchDatastore_Task, 'duration_secs': 0.008769} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.550479] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.550743] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.551070] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.551465] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.551584] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1ddadfd-f60c-4e03-a7d8-b2764a796bea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.554041] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73fd7e90-ec58-49cc-b4ca-70376351d33b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.562381] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 860.562381] env[61824]: value = "task-1275648" [ 860.562381] env[61824]: _type = "Task" [ 860.562381] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.571121] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275648, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.574864] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.575327] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.576258] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de20b8f6-0362-4588-adc1-f20beff03352 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.584998] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 860.584998] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d0a0d4-5f27-f2a3-91d0-51f888bc688e" [ 860.584998] env[61824]: _type = "Task" [ 860.584998] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.598860] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d0a0d4-5f27-f2a3-91d0-51f888bc688e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.603089] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275645, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.795584] env[61824]: DEBUG oslo_vmware.api [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275646, 'name': PowerOffVM_Task, 'duration_secs': 0.251126} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.795915] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.796109] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 860.796450] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f9297cb-8152-42cd-ae8b-181e6c3625aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.840302] env[61824]: DEBUG oslo_concurrency.lockutils [req-920b7503-a7c4-46f3-9f83-23fc0983259e req-b2142828-9e32-4b40-ac63-daf3c997e988 service nova] Releasing lock "refresh_cache-cabad7a6-206f-4e02-be3c-e18654396b7e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.877163] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 860.877540] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 860.877969] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Deleting the datastore file [datastore1] 37c405be-62a1-4feb-8cae-ea1c924f4f40 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.881904] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9aaddaa6-022d-4a07-8b62-ff850faf8700 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.886279] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275647, 'name': Rename_Task, 'duration_secs': 0.244535} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.886279] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.886279] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2aa48e36-0803-4e95-9e55-f62db65b7260 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.889588] env[61824]: DEBUG oslo_vmware.api [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for the task: (returnval){ [ 860.889588] env[61824]: value = "task-1275650" [ 860.889588] env[61824]: _type = "Task" [ 860.889588] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.895929] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 860.895929] env[61824]: value = "task-1275651" [ 860.895929] env[61824]: _type = "Task" [ 860.895929] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.903205] env[61824]: DEBUG oslo_vmware.api [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275650, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.911098] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275651, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.989265] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.989668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.013836] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275644, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.075201] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275648, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.099098] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d0a0d4-5f27-f2a3-91d0-51f888bc688e, 'name': SearchDatastore_Task, 'duration_secs': 0.084091} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.103282] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275645, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.103579] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04a94bd1-e954-4767-a1ec-d2167db71c9a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.110487] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 861.110487] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a75866-51ab-c3ca-ec74-3d21e42bc6de" [ 861.110487] env[61824]: _type = "Task" [ 861.110487] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.121994] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a75866-51ab-c3ca-ec74-3d21e42bc6de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.343586] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.687s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.347379] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.585s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.349357] env[61824]: INFO nova.compute.claims [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.406382] env[61824]: DEBUG oslo_vmware.api [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Task: {'id': task-1275650, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26369} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.407156] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.407376] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 861.407591] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 861.407791] env[61824]: INFO nova.compute.manager [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Took 1.14 seconds to destroy the instance on the hypervisor. [ 861.408103] env[61824]: DEBUG oslo.service.loopingcall [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.408381] env[61824]: DEBUG nova.compute.manager [-] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 861.408554] env[61824]: DEBUG nova.network.neutron [-] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 861.414920] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275651, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.493968] env[61824]: DEBUG nova.compute.utils [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.507266] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275644, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.520404] env[61824]: DEBUG nova.objects.instance [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lazy-loading 'flavor' on Instance uuid d9bb3172-c058-4e59-a719-d8aa340fba71 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.572971] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275648, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.595648] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275645, 'name': CloneVM_Task} progress is 95%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.621738] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a75866-51ab-c3ca-ec74-3d21e42bc6de, 'name': SearchDatastore_Task, 'duration_secs': 0.104644} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.622188] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.622514] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] cabad7a6-206f-4e02-be3c-e18654396b7e/cabad7a6-206f-4e02-be3c-e18654396b7e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.622869] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-81ce9287-063a-43eb-b24c-1d6c3c60dfcb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.630250] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 861.630250] env[61824]: value = "task-1275652" [ 861.630250] env[61824]: _type = "Task" [ 861.630250] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.639670] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.908959] env[61824]: DEBUG oslo_vmware.api [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275651, 'name': PowerOnVM_Task, 'duration_secs': 0.64585} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.909505] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.909893] env[61824]: INFO nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Took 11.76 seconds to spawn the instance on the hypervisor. [ 861.910259] env[61824]: DEBUG nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.911560] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20535448-690b-44ef-bb33-640ba199b8c2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.922214] env[61824]: INFO nova.scheduler.client.report [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted allocation for migration de82b366-9a8e-489e-bd6f-2729751101dc [ 861.983841] env[61824]: DEBUG nova.compute.manager [req-c20e4f20-d113-494f-9741-d2d4f8310321 req-d4aa77f6-63e5-4f67-b100-bb6af722d26e service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Received event network-vif-deleted-beb8cba8-71b3-40a5-ba5d-665a9770ec4d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.984112] env[61824]: INFO nova.compute.manager [req-c20e4f20-d113-494f-9741-d2d4f8310321 req-d4aa77f6-63e5-4f67-b100-bb6af722d26e service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Neutron deleted interface beb8cba8-71b3-40a5-ba5d-665a9770ec4d; detaching it from the instance and deleting it from the info cache [ 861.984262] env[61824]: DEBUG nova.network.neutron [req-c20e4f20-d113-494f-9741-d2d4f8310321 req-d4aa77f6-63e5-4f67-b100-bb6af722d26e service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.001710] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.010351] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275644, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.026602] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa3e4dc8-59cf-469b-ae55-669213ee20ae tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.430s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.073082] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275648, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.095255] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275645, 'name': CloneVM_Task, 'duration_secs': 1.710475} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.095548] env[61824]: INFO nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Created linked-clone VM from snapshot [ 862.096941] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb089c52-7d2e-4dcd-9640-cfb9540e46f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.103735] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Uploading image 250a0014-d97e-44f5-8fee-cc58ffff714d {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 862.137292] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 862.137292] env[61824]: value = "vm-274170" [ 862.137292] env[61824]: _type = "VirtualMachine" [ 862.137292] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 862.137615] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bc21e9fc-42ed-425f-be91-e7a22b4ff2a5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.145059] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.147034] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease: (returnval){ [ 862.147034] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525acc55-ac84-cbec-5f81-ebf6c25d9d16" [ 862.147034] env[61824]: _type = "HttpNfcLease" [ 862.147034] env[61824]: } obtained for exporting VM: (result){ [ 862.147034] env[61824]: value = "vm-274170" [ 862.147034] env[61824]: _type = "VirtualMachine" [ 862.147034] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 862.147491] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the lease: (returnval){ [ 862.147491] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525acc55-ac84-cbec-5f81-ebf6c25d9d16" [ 862.147491] env[61824]: _type = "HttpNfcLease" [ 862.147491] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 862.155247] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 862.155247] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525acc55-ac84-cbec-5f81-ebf6c25d9d16" [ 862.155247] env[61824]: _type = "HttpNfcLease" [ 862.155247] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 862.235452] env[61824]: DEBUG nova.network.neutron [-] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.343784] env[61824]: DEBUG oslo_concurrency.lockutils [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "d9bb3172-c058-4e59-a719-d8aa340fba71" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.343947] env[61824]: DEBUG oslo_concurrency.lockutils [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.427980] env[61824]: DEBUG oslo_concurrency.lockutils [None req-265a7a6d-a8a5-4cc9-8d46-0b2bf31a7583 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 16.333s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.438657] env[61824]: INFO nova.compute.manager [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Took 21.38 seconds to build instance. [ 862.489103] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3efa2025-c0e3-4842-ab48-987bda744ed9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.506769] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101867f8-ebbf-49ca-9bda-62dfbb7721f4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.528449] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275644, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.549402] env[61824]: DEBUG nova.compute.manager [req-c20e4f20-d113-494f-9741-d2d4f8310321 req-d4aa77f6-63e5-4f67-b100-bb6af722d26e service nova] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Detach interface failed, port_id=beb8cba8-71b3-40a5-ba5d-665a9770ec4d, reason: Instance 37c405be-62a1-4feb-8cae-ea1c924f4f40 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 862.578842] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275648, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.643544] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.656143] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 862.656143] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525acc55-ac84-cbec-5f81-ebf6c25d9d16" [ 862.656143] env[61824]: _type = "HttpNfcLease" [ 862.656143] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 862.656454] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 862.656454] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525acc55-ac84-cbec-5f81-ebf6c25d9d16" [ 862.656454] env[61824]: _type = "HttpNfcLease" [ 862.656454] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 862.657229] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bcd00dc-22ca-4eaa-b2cb-f968ed36fa6c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.668794] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5212a4dc-839f-075e-c243-c8f038b5cb27/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 862.669240] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5212a4dc-839f-075e-c243-c8f038b5cb27/disk-0.vmdk for reading. {{(pid=61824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 862.739092] env[61824]: INFO nova.compute.manager [-] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Took 1.33 seconds to deallocate network for instance. [ 862.766048] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-060d5960-5e0d-4472-bb9f-2765699bb1de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.775039] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eabed4-d878-4f4c-92b1-d5f7034338d9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.782798] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a15965-14b2-496e-842b-37ad54ac2b0b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.823784] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9526ca3a-cab1-46cb-ba93-3c34ad37f4af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.844591] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae3fd55-f724-4e58-8f65-1f0705d5fe89 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.853217] env[61824]: INFO nova.compute.manager [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Detaching volume 6c0ac2ed-db17-4169-b133-158c17007c69 [ 862.872589] env[61824]: DEBUG nova.compute.provider_tree [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.918679] env[61824]: INFO nova.virt.block_device [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Attempting to driver detach volume 6c0ac2ed-db17-4169-b133-158c17007c69 from mountpoint /dev/sdb [ 862.919053] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 862.919210] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274163', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'name': 'volume-6c0ac2ed-db17-4169-b133-158c17007c69', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9bb3172-c058-4e59-a719-d8aa340fba71', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'serial': '6c0ac2ed-db17-4169-b133-158c17007c69'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 862.920131] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a097292d-6dc8-421d-9720-6a6e38ac84fe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.942676] env[61824]: DEBUG oslo_concurrency.lockutils [None req-503f8177-d047-4a3e-a0ed-48e7f10ef1ac tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.897s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.943604] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1395ed5-2e2b-4989-a533-552c335713cb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.951257] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9880f7f-78b4-44af-ba8c-254c5a32041f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.973951] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946ed55f-056b-4572-93a5-d8c6d2a7b505 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.991670] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] The volume has not been displaced from its original location: [datastore2] volume-6c0ac2ed-db17-4169-b133-158c17007c69/volume-6c0ac2ed-db17-4169-b133-158c17007c69.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 862.997517] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Reconfiguring VM instance instance-0000003f to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 862.998018] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dbb3c577-7fe1-4dfd-a34f-9d98b7e7e595 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.022168] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275644, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.692475} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.023655] env[61824]: INFO nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4/OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4.vmdk to [datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650/619a0c61-29bb-40c8-bc1c-82d0621b6650.vmdk. [ 863.023783] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Cleaning up location [datastore2] OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 863.023936] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_efbd0cdf-135e-457d-9868-9901534922a4 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.024278] env[61824]: DEBUG oslo_vmware.api [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 863.024278] env[61824]: value = "task-1275654" [ 863.024278] env[61824]: _type = "Task" [ 863.024278] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.024471] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e0737df-953e-493c-b2a0-2439ee1c6e46 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.034492] env[61824]: DEBUG oslo_vmware.api [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275654, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.037324] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 863.037324] env[61824]: value = "task-1275655" [ 863.037324] env[61824]: _type = "Task" [ 863.037324] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.047306] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275655, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.065951] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.066238] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.066471] env[61824]: INFO nova.compute.manager [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Attaching volume 779f9475-aa56-4fae-b35f-ce65c3cfc4be to /dev/sdb [ 863.077668] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275648, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.129266} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.077970] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.078219] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.078490] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f5876ab-8391-4e29-8278-09e1933fdff9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.084498] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 863.084498] env[61824]: value = "task-1275656" [ 863.084498] env[61824]: _type = "Task" [ 863.084498] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.092583] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.111869] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868d18b4-e040-4a7e-8378-5a8c475daced {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.120090] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d74d09-b465-49fa-b427-fd64451dc2f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.137196] env[61824]: DEBUG nova.virt.block_device [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updating existing volume attachment record: 289ebd93-39ce-429c-b3ff-62aaa3d17354 {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 863.148336] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275652, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.40608} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.148616] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] cabad7a6-206f-4e02-be3c-e18654396b7e/cabad7a6-206f-4e02-be3c-e18654396b7e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.148826] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.149122] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-527340bc-5ef9-422d-b2d6-30e41499e68b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.155870] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 863.155870] env[61824]: value = "task-1275657" [ 863.155870] env[61824]: _type = "Task" [ 863.155870] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.163878] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275657, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.251789] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.379269] env[61824]: DEBUG nova.scheduler.client.report [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.537100] env[61824]: DEBUG oslo_vmware.api [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275654, 'name': ReconfigVM_Task, 'duration_secs': 0.46235} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.541145] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Reconfigured VM instance instance-0000003f to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 863.544498] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e3ea020-bbee-45ce-89c3-002178d7887f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.564962] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275655, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040063} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.566594] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.566942] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650/619a0c61-29bb-40c8-bc1c-82d0621b6650.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.567528] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650/619a0c61-29bb-40c8-bc1c-82d0621b6650.vmdk to [datastore2] 5806faad-2bab-4fda-a131-566af50c4d7e/5806faad-2bab-4fda-a131-566af50c4d7e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.568099] env[61824]: DEBUG oslo_vmware.api [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 863.568099] env[61824]: value = "task-1275661" [ 863.568099] env[61824]: _type = "Task" [ 863.568099] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.568519] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3285136b-9c24-494e-bda0-f19288e0d56a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.580245] env[61824]: DEBUG oslo_vmware.api [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275661, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.581940] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 863.581940] env[61824]: value = "task-1275662" [ 863.581940] env[61824]: _type = "Task" [ 863.581940] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.595368] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275662, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.601582] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060411} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.602264] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.603246] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59739505-3bfb-46a7-b854-baafa06a9024 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.628916] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.630140] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24828f3b-0861-4517-972b-b83b3392263c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.653590] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 863.653590] env[61824]: value = "task-1275663" [ 863.653590] env[61824]: _type = "Task" [ 863.653590] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.667388] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275663, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.669803] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275657, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068559} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.670366] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.671345] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd1a35d-09b1-4b6d-bbed-3799bb75b7f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.696479] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] cabad7a6-206f-4e02-be3c-e18654396b7e/cabad7a6-206f-4e02-be3c-e18654396b7e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.696850] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4e351b1-58ce-41c8-9b7d-ee04f19175c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.720290] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 863.720290] env[61824]: value = "task-1275664" [ 863.720290] env[61824]: _type = "Task" [ 863.720290] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.730354] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275664, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.884575] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.885295] env[61824]: DEBUG nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 863.888731] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.820s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.889012] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.891538] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.699s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.893157] env[61824]: INFO nova.compute.claims [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.935827] env[61824]: INFO nova.scheduler.client.report [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Deleted allocations for instance 6235a9b9-19c0-4917-938a-bca82a2ca05e [ 864.081586] env[61824]: DEBUG oslo_vmware.api [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275661, 'name': ReconfigVM_Task, 'duration_secs': 0.235857} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.082224] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274163', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'name': 'volume-6c0ac2ed-db17-4169-b133-158c17007c69', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd9bb3172-c058-4e59-a719-d8aa340fba71', 'attached_at': '', 'detached_at': '', 'volume_id': '6c0ac2ed-db17-4169-b133-158c17007c69', 'serial': '6c0ac2ed-db17-4169-b133-158c17007c69'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 864.097128] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275662, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.165755] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275663, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.179833] env[61824]: DEBUG nova.compute.manager [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.180386] env[61824]: DEBUG nova.compute.manager [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing instance network info cache due to event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.180532] env[61824]: DEBUG oslo_concurrency.lockutils [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.180827] env[61824]: DEBUG oslo_concurrency.lockutils [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.181115] env[61824]: DEBUG nova.network.neutron [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 864.234860] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275664, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.398518] env[61824]: DEBUG nova.compute.utils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 864.403803] env[61824]: DEBUG nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 864.403803] env[61824]: DEBUG nova.network.neutron [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 864.451365] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9537dc2a-24ae-4efa-8f62-366e8adbd67f tempest-ServersTestBootFromVolume-160752249 tempest-ServersTestBootFromVolume-160752249-project-member] Lock "6235a9b9-19c0-4917-938a-bca82a2ca05e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.260s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.531803] env[61824]: DEBUG nova.policy [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f8e7019b3c754b079d0df72b0312d6f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc352363e7f94dc6b7a99db8aca565a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.599472] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275662, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.640873] env[61824]: DEBUG nova.objects.instance [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lazy-loading 'flavor' on Instance uuid d9bb3172-c058-4e59-a719-d8aa340fba71 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.673241] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275663, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.735903] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275664, 'name': ReconfigVM_Task, 'duration_secs': 0.519045} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.735903] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Reconfigured VM instance instance-0000004d to attach disk [datastore2] cabad7a6-206f-4e02-be3c-e18654396b7e/cabad7a6-206f-4e02-be3c-e18654396b7e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.735903] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-80cfb411-c39a-4d36-9491-bc4d76731b18 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.745734] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 864.745734] env[61824]: value = "task-1275665" [ 864.745734] env[61824]: _type = "Task" [ 864.745734] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.758487] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.758487] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.758487] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.758487] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.758978] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.760262] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275665, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.760933] env[61824]: INFO nova.compute.manager [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Terminating instance [ 864.763571] env[61824]: DEBUG nova.compute.manager [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 864.764084] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 864.765576] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e9d31e-8e9a-4a48-8292-319d2b3ea771 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.780431] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.781015] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f80612c-21ad-4714-abe7-6a14e99db9ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.789404] env[61824]: DEBUG oslo_vmware.api [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 864.789404] env[61824]: value = "task-1275666" [ 864.789404] env[61824]: _type = "Task" [ 864.789404] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.808420] env[61824]: DEBUG oslo_vmware.api [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275666, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.918041] env[61824]: DEBUG nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 864.979680] env[61824]: DEBUG nova.network.neutron [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updated VIF entry in instance network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.980308] env[61824]: DEBUG nova.network.neutron [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.107352] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275662, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.174126] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275663, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.177902] env[61824]: DEBUG nova.network.neutron [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Successfully created port: 66cc8af8-5b27-44c8-b043-106fe1e38714 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.263155] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275665, 'name': Rename_Task, 'duration_secs': 0.200757} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.264115] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.264453] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48e6978e-c003-4b01-a2e9-cbce4847eaea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.276073] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 865.276073] env[61824]: value = "task-1275667" [ 865.276073] env[61824]: _type = "Task" [ 865.276073] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.286295] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275667, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.300212] env[61824]: DEBUG oslo_vmware.api [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275666, 'name': PowerOffVM_Task, 'duration_secs': 0.289651} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.300737] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 865.301468] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 865.301468] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53b8dd16-f864-42c6-b3a6-a1eea4df3076 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.319991] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d51cc22-c873-4787-b3ee-e8090bb169b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.330944] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de08449-8b0a-4438-9463-85e314329543 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.370967] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8f7d7e-0c0c-41f9-93cd-ec1713a92de5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.382300] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b629ff3b-7b3e-4812-a73e-84fb30857dff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.388939] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 865.388939] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 865.388939] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleting the datastore file [datastore1] a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.388939] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88d61b3e-e7d1-451c-b7b0-4f044b712127 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.401090] env[61824]: DEBUG nova.compute.provider_tree [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.403912] env[61824]: DEBUG oslo_vmware.api [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 865.403912] env[61824]: value = "task-1275669" [ 865.403912] env[61824]: _type = "Task" [ 865.403912] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.414700] env[61824]: DEBUG oslo_vmware.api [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275669, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.490657] env[61824]: DEBUG oslo_concurrency.lockutils [req-26ae6964-d328-4c81-8fbe-19e654fd6724 req-de099f90-b050-4178-af27-650822c081f6 service nova] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.606954] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "09432e72-c10a-4a7a-93af-af0d371fb871" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.607497] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.609410] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275662, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.664977] env[61824]: DEBUG oslo_concurrency.lockutils [None req-81624c80-7cea-4032-87f9-310253c3b675 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.321s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.675065] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275663, 'name': ReconfigVM_Task, 'duration_secs': 1.705539} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.675448] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.676457] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-122fe074-5494-401f-9a90-fe01fd1b8b6c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.686828] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 865.686828] env[61824]: value = "task-1275671" [ 865.686828] env[61824]: _type = "Task" [ 865.686828] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.699735] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275671, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.788953] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275667, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.909359] env[61824]: DEBUG nova.scheduler.client.report [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.921465] env[61824]: DEBUG oslo_vmware.api [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275669, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291651} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.921465] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 865.921465] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 865.921465] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 865.921465] env[61824]: INFO nova.compute.manager [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Took 1.16 seconds to destroy the instance on the hypervisor. [ 865.921811] env[61824]: DEBUG oslo.service.loopingcall [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.921811] env[61824]: DEBUG nova.compute.manager [-] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 865.921811] env[61824]: DEBUG nova.network.neutron [-] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 865.933973] env[61824]: DEBUG nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 865.967590] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.968261] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.968261] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.968391] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.968426] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.969298] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.969298] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.969298] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.969298] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.969298] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.969495] env[61824]: DEBUG nova.virt.hardware [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.970716] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db18e49-57f3-455a-884c-5547d852f4cf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.982237] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7790396d-0c9c-47cf-9f6e-73006e412be7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.099090] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275662, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.497911} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.099384] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/619a0c61-29bb-40c8-bc1c-82d0621b6650/619a0c61-29bb-40c8-bc1c-82d0621b6650.vmdk to [datastore2] 5806faad-2bab-4fda-a131-566af50c4d7e/5806faad-2bab-4fda-a131-566af50c4d7e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.100285] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4197c2d-900c-4f78-b235-71ab0adc5091 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.118175] env[61824]: DEBUG nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.134390] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 5806faad-2bab-4fda-a131-566af50c4d7e/5806faad-2bab-4fda-a131-566af50c4d7e.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.138285] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e63e521-a895-4b45-a9a0-435f598017aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.156665] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 866.156665] env[61824]: value = "task-1275672" [ 866.156665] env[61824]: _type = "Task" [ 866.156665] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.166931] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275672, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.197720] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275671, 'name': Rename_Task, 'duration_secs': 0.467572} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.198090] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.198939] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63a6149b-bb1b-44ba-b14f-555e4131e062 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.204909] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 866.204909] env[61824]: value = "task-1275673" [ 866.204909] env[61824]: _type = "Task" [ 866.204909] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.213558] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.238144] env[61824]: DEBUG nova.compute.manager [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.238431] env[61824]: DEBUG nova.compute.manager [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing instance network info cache due to event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.238666] env[61824]: DEBUG oslo_concurrency.lockutils [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.238857] env[61824]: DEBUG oslo_concurrency.lockutils [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.239313] env[61824]: DEBUG nova.network.neutron [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing network info cache for port bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 866.288039] env[61824]: DEBUG oslo_vmware.api [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275667, 'name': PowerOnVM_Task, 'duration_secs': 0.595648} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.288039] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.288039] env[61824]: INFO nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Took 9.11 seconds to spawn the instance on the hypervisor. [ 866.288039] env[61824]: DEBUG nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.288326] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061f390f-0274-4114-ac26-87cd3f4d435c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.414795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.523s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.415771] env[61824]: DEBUG nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 866.417973] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.015s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.418227] env[61824]: DEBUG nova.objects.instance [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lazy-loading 'resources' on Instance uuid 31edad40-a743-4b65-8ef0-615a9ba2a225 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.564542] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "d9bb3172-c058-4e59-a719-d8aa340fba71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.564806] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.565121] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "d9bb3172-c058-4e59-a719-d8aa340fba71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.565322] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.565520] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.568242] env[61824]: INFO nova.compute.manager [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Terminating instance [ 866.572660] env[61824]: DEBUG nova.compute.manager [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 866.572857] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.573734] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0ef0b1-397b-475a-a6ab-c0eaeb4d21d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.582845] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.583133] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01f7f952-477e-4dec-9c7b-05d1dfc76b26 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.589461] env[61824]: DEBUG oslo_vmware.api [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 866.589461] env[61824]: value = "task-1275674" [ 866.589461] env[61824]: _type = "Task" [ 866.589461] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.597868] env[61824]: DEBUG oslo_vmware.api [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.657706] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.669572] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275672, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.715626] env[61824]: DEBUG oslo_vmware.api [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275673, 'name': PowerOnVM_Task, 'duration_secs': 0.454917} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.715996] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.716307] env[61824]: DEBUG nova.compute.manager [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 866.717229] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a84284-e87d-47f2-97e4-51e3084fe02f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.810466] env[61824]: INFO nova.compute.manager [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Took 18.93 seconds to build instance. [ 866.926488] env[61824]: DEBUG nova.compute.utils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.930769] env[61824]: DEBUG nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.930950] env[61824]: DEBUG nova.network.neutron [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 867.034405] env[61824]: DEBUG nova.policy [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '59cac3d5f8874eb080b327942cbd6798', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8b302783679441b7bc76b291562031d7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.046587] env[61824]: DEBUG nova.compute.manager [req-3b8e8233-9d65-41fe-8d0f-5f7d4af5d298 req-ac15a9ad-f4fa-4141-88ef-e54256409118 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Received event network-vif-deleted-e95afb5b-11f8-4ed4-abb0-a12b48c9c14d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.046867] env[61824]: INFO nova.compute.manager [req-3b8e8233-9d65-41fe-8d0f-5f7d4af5d298 req-ac15a9ad-f4fa-4141-88ef-e54256409118 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Neutron deleted interface e95afb5b-11f8-4ed4-abb0-a12b48c9c14d; detaching it from the instance and deleting it from the info cache [ 867.047108] env[61824]: DEBUG nova.network.neutron [req-3b8e8233-9d65-41fe-8d0f-5f7d4af5d298 req-ac15a9ad-f4fa-4141-88ef-e54256409118 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.099813] env[61824]: DEBUG oslo_vmware.api [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275674, 'name': PowerOffVM_Task, 'duration_secs': 0.379104} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.100097] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.100287] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 867.100545] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5332ad9-7187-4d99-855f-d302ff4c437f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.148015] env[61824]: DEBUG nova.network.neutron [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updated VIF entry in instance network info cache for port bd35af69-df49-4c0a-a798-57bd9c8f6fe4. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.148442] env[61824]: DEBUG nova.network.neutron [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.169176] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275672, 'name': ReconfigVM_Task, 'duration_secs': 0.82825} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.171998] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 5806faad-2bab-4fda-a131-566af50c4d7e/5806faad-2bab-4fda-a131-566af50c4d7e.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.173835] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89b36e62-c79a-4c44-b704-143d0e877d39 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.175877] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 867.176195] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 867.176336] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Deleting the datastore file [datastore2] d9bb3172-c058-4e59-a719-d8aa340fba71 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.176585] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75948fa3-5b69-44e5-a56a-85caf6aef9ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.183266] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 867.183266] env[61824]: value = "task-1275676" [ 867.183266] env[61824]: _type = "Task" [ 867.183266] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.184946] env[61824]: DEBUG oslo_vmware.api [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 867.184946] env[61824]: value = "task-1275677" [ 867.184946] env[61824]: _type = "Task" [ 867.184946] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.201241] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275676, 'name': Rename_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.202836] env[61824]: DEBUG oslo_vmware.api [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275677, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.234017] env[61824]: DEBUG nova.network.neutron [-] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.242273] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.311090] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9f6ff1-f578-4417-b36c-67a89273a974 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.314746] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ce419de8-7e6c-4aed-8634-0c38bb624562 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.447s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.321086] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8673efd6-e9e7-4c9f-a58d-9defdc29f04d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.354912] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00ba21ac-85e1-46fa-b9a0-dfc1757b3cd3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.363508] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedfd9e2-a736-4b99-b3ff-78b7c8b93181 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.379799] env[61824]: DEBUG nova.compute.provider_tree [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.432100] env[61824]: DEBUG nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.519694] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "cabad7a6-206f-4e02-be3c-e18654396b7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.519998] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.520306] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "cabad7a6-206f-4e02-be3c-e18654396b7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.520512] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.520712] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.529605] env[61824]: INFO nova.compute.manager [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Terminating instance [ 867.532190] env[61824]: DEBUG nova.compute.manager [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 867.532277] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 867.533080] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1ceb2c-38be-4c93-9a74-adb1e45d7c86 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.546247] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 867.548379] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a4e7c0b-bfd5-4e7a-bec3-71b729bf9563 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.552543] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08777a7f-bb85-4394-acb5-c807d98a3d56 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.558019] env[61824]: DEBUG oslo_vmware.api [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 867.558019] env[61824]: value = "task-1275678" [ 867.558019] env[61824]: _type = "Task" [ 867.558019] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.571083] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03fe39c-9ba4-47b3-9415-756d1a9b2915 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.582987] env[61824]: DEBUG oslo_vmware.api [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275678, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.605321] env[61824]: DEBUG nova.compute.manager [req-3b8e8233-9d65-41fe-8d0f-5f7d4af5d298 req-ac15a9ad-f4fa-4141-88ef-e54256409118 service nova] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Detach interface failed, port_id=e95afb5b-11f8-4ed4-abb0-a12b48c9c14d, reason: Instance a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 867.654275] env[61824]: DEBUG oslo_concurrency.lockutils [req-5dec18eb-01ad-47d7-a60f-a5499ceb8413 req-3faeca73-8e0a-4288-a55b-0e932fca2765 service nova] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.700710] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 867.701141] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274172', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'name': 'volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'serial': '779f9475-aa56-4fae-b35f-ce65c3cfc4be'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 867.702687] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b025f84-6b17-42c1-9afe-68e5c031fa83 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.712298] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275676, 'name': Rename_Task, 'duration_secs': 0.215647} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.712876] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.716365] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05f5f690-ade4-4dbf-9e8c-544b160cdd00 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.719038] env[61824]: DEBUG oslo_vmware.api [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275677, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280403} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.733747] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.734081] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 867.734175] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 867.734354] env[61824]: INFO nova.compute.manager [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Took 1.16 seconds to destroy the instance on the hypervisor. [ 867.734605] env[61824]: DEBUG oslo.service.loopingcall [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.736044] env[61824]: INFO nova.compute.manager [-] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Took 1.81 seconds to deallocate network for instance. [ 867.736298] env[61824]: DEBUG nova.compute.manager [-] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.736390] env[61824]: DEBUG nova.network.neutron [-] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 867.740548] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a68a49-f260-4636-8dd5-35de333a7fee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.748383] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 867.748383] env[61824]: value = "task-1275679" [ 867.748383] env[61824]: _type = "Task" [ 867.748383] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.782700] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be/volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.784212] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1280d380-5452-4316-9139-91cc3740b77a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.804588] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275679, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.812972] env[61824]: DEBUG oslo_vmware.api [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 867.812972] env[61824]: value = "task-1275680" [ 867.812972] env[61824]: _type = "Task" [ 867.812972] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.812972] env[61824]: DEBUG nova.network.neutron [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Successfully updated port: 66cc8af8-5b27-44c8-b043-106fe1e38714 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.827255] env[61824]: DEBUG oslo_vmware.api [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275680, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.868805] env[61824]: DEBUG nova.network.neutron [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Successfully created port: 1a99ec03-ba39-4515-827d-6c5c7781f603 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 867.883935] env[61824]: DEBUG nova.scheduler.client.report [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.070075] env[61824]: DEBUG oslo_vmware.api [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275678, 'name': PowerOffVM_Task, 'duration_secs': 0.331369} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.070410] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 868.070587] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 868.070855] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b1c6d5c-03be-4e10-9286-0cf6e6721c23 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.134218] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 868.134589] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 868.135128] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleting the datastore file [datastore2] cabad7a6-206f-4e02-be3c-e18654396b7e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.135545] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-540ee8aa-6edb-44e3-be54-0bb7355a2de1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.145149] env[61824]: DEBUG oslo_vmware.api [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 868.145149] env[61824]: value = "task-1275682" [ 868.145149] env[61824]: _type = "Task" [ 868.145149] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.153111] env[61824]: DEBUG oslo_vmware.api [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275682, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.252612] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.267264] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275679, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.285560] env[61824]: DEBUG nova.compute.manager [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Received event network-vif-plugged-66cc8af8-5b27-44c8-b043-106fe1e38714 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.285560] env[61824]: DEBUG oslo_concurrency.lockutils [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] Acquiring lock "c6552176-1128-473b-ac4a-f3b55428fd9e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.285560] env[61824]: DEBUG oslo_concurrency.lockutils [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.285560] env[61824]: DEBUG oslo_concurrency.lockutils [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.285560] env[61824]: DEBUG nova.compute.manager [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] No waiting events found dispatching network-vif-plugged-66cc8af8-5b27-44c8-b043-106fe1e38714 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 868.285896] env[61824]: WARNING nova.compute.manager [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Received unexpected event network-vif-plugged-66cc8af8-5b27-44c8-b043-106fe1e38714 for instance with vm_state building and task_state spawning. [ 868.285896] env[61824]: DEBUG nova.compute.manager [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Received event network-changed-66cc8af8-5b27-44c8-b043-106fe1e38714 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.285896] env[61824]: DEBUG nova.compute.manager [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Refreshing instance network info cache due to event network-changed-66cc8af8-5b27-44c8-b043-106fe1e38714. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 868.285896] env[61824]: DEBUG oslo_concurrency.lockutils [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] Acquiring lock "refresh_cache-c6552176-1128-473b-ac4a-f3b55428fd9e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.286262] env[61824]: DEBUG oslo_concurrency.lockutils [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] Acquired lock "refresh_cache-c6552176-1128-473b-ac4a-f3b55428fd9e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.286262] env[61824]: DEBUG nova.network.neutron [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Refreshing network info cache for port 66cc8af8-5b27-44c8-b043-106fe1e38714 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 868.321180] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "refresh_cache-c6552176-1128-473b-ac4a-f3b55428fd9e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.326622] env[61824]: DEBUG oslo_vmware.api [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275680, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.389913] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.972s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.393772] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.142s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.394126] env[61824]: DEBUG nova.objects.instance [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lazy-loading 'resources' on Instance uuid 37c405be-62a1-4feb-8cae-ea1c924f4f40 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 868.420373] env[61824]: INFO nova.scheduler.client.report [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Deleted allocations for instance 31edad40-a743-4b65-8ef0-615a9ba2a225 [ 868.441200] env[61824]: DEBUG nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 868.482637] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.483311] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.483553] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.485082] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.485082] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.485082] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.485082] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.485082] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.485420] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.485420] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.485420] env[61824]: DEBUG nova.virt.hardware [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.487220] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e6e3db-ae49-46a3-96f0-c0ae5bbd4449 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.498057] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41de967-4e32-40a8-bb7d-7548b13947de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.653874] env[61824]: DEBUG oslo_vmware.api [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275682, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.273555} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.654275] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.654914] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 868.655407] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 868.655644] env[61824]: INFO nova.compute.manager [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 868.656014] env[61824]: DEBUG oslo.service.loopingcall [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.656631] env[61824]: DEBUG nova.compute.manager [-] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 868.656706] env[61824]: DEBUG nova.network.neutron [-] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.763630] env[61824]: DEBUG oslo_vmware.api [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275679, 'name': PowerOnVM_Task, 'duration_secs': 0.620508} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.763919] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.764188] env[61824]: INFO nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Took 14.02 seconds to spawn the instance on the hypervisor. [ 868.764412] env[61824]: DEBUG nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.765421] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ed923d-7089-4c1c-aedb-cb7f4aa661ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.832963] env[61824]: DEBUG oslo_vmware.api [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275680, 'name': ReconfigVM_Task, 'duration_secs': 0.543135} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.833608] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfigured VM instance instance-00000042 to attach disk [datastore1] volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be/volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.845130] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef00bd5a-c24e-4cc0-a0c4-99050f001ca4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.866192] env[61824]: DEBUG oslo_vmware.api [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 868.866192] env[61824]: value = "task-1275683" [ 868.866192] env[61824]: _type = "Task" [ 868.866192] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.879031] env[61824]: DEBUG oslo_vmware.api [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275683, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.901522] env[61824]: DEBUG nova.network.neutron [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 868.933494] env[61824]: DEBUG oslo_concurrency.lockutils [None req-342a29ef-d7cf-4673-bb52-007fe4038168 tempest-ServersNegativeTestMultiTenantJSON-1989973855 tempest-ServersNegativeTestMultiTenantJSON-1989973855-project-member] Lock "31edad40-a743-4b65-8ef0-615a9ba2a225" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.471s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.015152] env[61824]: DEBUG nova.network.neutron [-] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.069132] env[61824]: DEBUG nova.network.neutron [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.144949] env[61824]: DEBUG nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.144949] env[61824]: DEBUG nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing instance network info cache due to event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 869.144949] env[61824]: DEBUG oslo_concurrency.lockutils [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.144949] env[61824]: DEBUG oslo_concurrency.lockutils [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.145343] env[61824]: DEBUG nova.network.neutron [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing network info cache for port bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 869.229777] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c56ce50-503a-420c-89ba-2dd1ecb2beaa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.239088] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdb758d-095f-43ee-bec3-5744edf1590b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.271266] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413f96b0-cf89-4d82-8c7b-5ebc963797e5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.283683] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc93d4b8-9fe2-4cdf-93af-bf7c52f0599b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.291403] env[61824]: INFO nova.compute.manager [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Took 24.65 seconds to build instance. [ 869.294234] env[61824]: INFO nova.compute.manager [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Rebuilding instance [ 869.305805] env[61824]: DEBUG nova.compute.provider_tree [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.354570] env[61824]: DEBUG nova.compute.manager [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.355884] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aae650c-8f66-45ba-b5ed-ffdfdafdb027 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.378534] env[61824]: DEBUG oslo_vmware.api [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275683, 'name': ReconfigVM_Task, 'duration_secs': 0.150946} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.379277] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274172', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'name': 'volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'serial': '779f9475-aa56-4fae-b35f-ce65c3cfc4be'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 869.480827] env[61824]: DEBUG nova.network.neutron [-] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.521300] env[61824]: INFO nova.compute.manager [-] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Took 1.78 seconds to deallocate network for instance. [ 869.571360] env[61824]: DEBUG oslo_concurrency.lockutils [req-8ebc02aa-43a2-4431-9e6c-36f4c8934399 req-d1c4ba23-4950-43f7-bfbd-2cf508dd6120 service nova] Releasing lock "refresh_cache-c6552176-1128-473b-ac4a-f3b55428fd9e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.571688] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "refresh_cache-c6552176-1128-473b-ac4a-f3b55428fd9e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.577165] env[61824]: DEBUG nova.network.neutron [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.612703] env[61824]: DEBUG nova.network.neutron [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Successfully updated port: 1a99ec03-ba39-4515-827d-6c5c7781f603 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.793972] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd6dc7c2-cb57-400d-a9a3-f6eba528b3ca tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5806faad-2bab-4fda-a131-566af50c4d7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.161s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.808995] env[61824]: DEBUG nova.scheduler.client.report [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.868330] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.869086] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc9c4070-8102-4403-b3e9-36a0dde5173d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.880715] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 869.880715] env[61824]: value = "task-1275684" [ 869.880715] env[61824]: _type = "Task" [ 869.880715] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.900877] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275684, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.984369] env[61824]: INFO nova.compute.manager [-] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Took 1.33 seconds to deallocate network for instance. [ 870.030316] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.112356] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "refresh_cache-b1a770f9-71ea-4f80-a562-c08a313b5753" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.112537] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquired lock "refresh_cache-b1a770f9-71ea-4f80-a562-c08a313b5753" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.112672] env[61824]: DEBUG nova.network.neutron [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 870.127624] env[61824]: DEBUG nova.network.neutron [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updated VIF entry in instance network info cache for port bd35af69-df49-4c0a-a798-57bd9c8f6fe4. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 870.133183] env[61824]: DEBUG nova.network.neutron [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.152128] env[61824]: DEBUG nova.network.neutron [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 870.321489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.928s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.328025] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.667s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.328025] env[61824]: INFO nova.compute.claims [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.348075] env[61824]: DEBUG nova.compute.manager [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Received event network-vif-plugged-1a99ec03-ba39-4515-827d-6c5c7781f603 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.348325] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] Acquiring lock "b1a770f9-71ea-4f80-a562-c08a313b5753-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.348549] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.348723] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.348896] env[61824]: DEBUG nova.compute.manager [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] No waiting events found dispatching network-vif-plugged-1a99ec03-ba39-4515-827d-6c5c7781f603 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 870.349971] env[61824]: WARNING nova.compute.manager [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Received unexpected event network-vif-plugged-1a99ec03-ba39-4515-827d-6c5c7781f603 for instance with vm_state building and task_state spawning. [ 870.350860] env[61824]: DEBUG nova.compute.manager [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Received event network-changed-1a99ec03-ba39-4515-827d-6c5c7781f603 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.352302] env[61824]: DEBUG nova.compute.manager [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Refreshing instance network info cache due to event network-changed-1a99ec03-ba39-4515-827d-6c5c7781f603. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.352302] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] Acquiring lock "refresh_cache-b1a770f9-71ea-4f80-a562-c08a313b5753" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.357740] env[61824]: INFO nova.scheduler.client.report [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Deleted allocations for instance 37c405be-62a1-4feb-8cae-ea1c924f4f40 [ 870.365553] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "5806faad-2bab-4fda-a131-566af50c4d7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.368461] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5806faad-2bab-4fda-a131-566af50c4d7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.368461] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "5806faad-2bab-4fda-a131-566af50c4d7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.368461] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5806faad-2bab-4fda-a131-566af50c4d7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.368461] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5806faad-2bab-4fda-a131-566af50c4d7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.372033] env[61824]: INFO nova.compute.manager [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Terminating instance [ 870.374519] env[61824]: DEBUG nova.compute.manager [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 870.374616] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.376284] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cee75b2-49b5-46f6-98ff-a2f7d4adb318 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.386023] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.386023] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0c65415-b4ce-47fc-9c35-4b315d97a34d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.396924] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275684, 'name': PowerOffVM_Task, 'duration_secs': 0.201305} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.398523] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.398834] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.399279] env[61824]: DEBUG oslo_vmware.api [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 870.399279] env[61824]: value = "task-1275685" [ 870.399279] env[61824]: _type = "Task" [ 870.399279] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.400476] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3dc4ac-7cf6-4a98-90d1-fa8ce5f01ba6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.413975] env[61824]: DEBUG oslo_vmware.api [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.416903] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.417601] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d61d951-5f64-4bcf-8025-f74529208d8d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.429031] env[61824]: DEBUG nova.network.neutron [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Updating instance_info_cache with network_info: [{"id": "66cc8af8-5b27-44c8-b043-106fe1e38714", "address": "fa:16:3e:4c:bb:be", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66cc8af8-5b", "ovs_interfaceid": "66cc8af8-5b27-44c8-b043-106fe1e38714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.437063] env[61824]: DEBUG nova.objects.instance [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lazy-loading 'flavor' on Instance uuid cbded637-56b9-4048-b66e-1edf6311f435 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.489071] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.489071] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.489071] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleting the datastore file [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.489071] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92c88190-d363-4f5b-a1d1-1c8d4e6bf74d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.493130] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.495812] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 870.495812] env[61824]: value = "task-1275687" [ 870.495812] env[61824]: _type = "Task" [ 870.495812] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.509704] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275687, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.632021] env[61824]: DEBUG oslo_concurrency.lockutils [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.634216] env[61824]: DEBUG nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.634216] env[61824]: DEBUG nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing instance network info cache due to event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.634216] env[61824]: DEBUG oslo_concurrency.lockutils [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.634216] env[61824]: DEBUG oslo_concurrency.lockutils [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.634547] env[61824]: DEBUG nova.network.neutron [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 870.654829] env[61824]: DEBUG nova.network.neutron [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 870.871569] env[61824]: DEBUG oslo_concurrency.lockutils [None req-92fc3d92-1837-4b50-894a-af846e334d82 tempest-ServerAddressesNegativeTestJSON-1020218995 tempest-ServerAddressesNegativeTestJSON-1020218995-project-member] Lock "37c405be-62a1-4feb-8cae-ea1c924f4f40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.613s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.882120] env[61824]: DEBUG nova.network.neutron [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Updating instance_info_cache with network_info: [{"id": "1a99ec03-ba39-4515-827d-6c5c7781f603", "address": "fa:16:3e:3d:dd:da", "network": {"id": "a9d5808f-a48e-482b-8b48-0de699d1ce80", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1537303780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b302783679441b7bc76b291562031d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a99ec03-ba", "ovs_interfaceid": "1a99ec03-ba39-4515-827d-6c5c7781f603", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.914599] env[61824]: DEBUG oslo_vmware.api [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275685, 'name': PowerOffVM_Task, 'duration_secs': 0.193811} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.914927] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.915178] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.915473] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f53ca159-487e-440f-b114-0102916a14db {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.936362] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "refresh_cache-c6552176-1128-473b-ac4a-f3b55428fd9e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.936362] env[61824]: DEBUG nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Instance network_info: |[{"id": "66cc8af8-5b27-44c8-b043-106fe1e38714", "address": "fa:16:3e:4c:bb:be", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66cc8af8-5b", "ovs_interfaceid": "66cc8af8-5b27-44c8-b043-106fe1e38714", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.936632] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:bb:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '66cc8af8-5b27-44c8-b043-106fe1e38714', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.946042] env[61824]: DEBUG oslo.service.loopingcall [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.948713] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.949404] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91ce1bd9-d29a-4f8f-aab8-7ebc1ef07c6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.964403] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a6129b67-a0ea-414a-bd50-de482d7d13cc tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.898s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.970991] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.970991] env[61824]: value = "task-1275689" [ 870.970991] env[61824]: _type = "Task" [ 870.970991] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.985125] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275689, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.985591] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.985831] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.986016] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleting the datastore file [datastore2] 5806faad-2bab-4fda-a131-566af50c4d7e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.987649] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0837daae-878d-46b9-b5ac-e0860f4c8c65 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.993397] env[61824]: DEBUG oslo_vmware.api [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 870.993397] env[61824]: value = "task-1275690" [ 870.993397] env[61824]: _type = "Task" [ 870.993397] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.007034] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275687, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139863} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.009947] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.010146] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.010396] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.013025] env[61824]: DEBUG oslo_vmware.api [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.340802] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5212a4dc-839f-075e-c243-c8f038b5cb27/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 871.344379] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147f9a95-b344-48df-962b-2e9252e3f803 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.351082] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5212a4dc-839f-075e-c243-c8f038b5cb27/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 871.351262] env[61824]: ERROR oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5212a4dc-839f-075e-c243-c8f038b5cb27/disk-0.vmdk due to incomplete transfer. [ 871.351503] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c1ebb93a-b38a-45d9-86a8-38994e5206e5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.359263] env[61824]: DEBUG oslo_vmware.rw_handles [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5212a4dc-839f-075e-c243-c8f038b5cb27/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 871.359445] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Uploaded image 250a0014-d97e-44f5-8fee-cc58ffff714d to the Glance image server {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 871.361888] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Destroying the VM {{(pid=61824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 871.362157] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fb0b8f5f-4210-4998-8f4d-7b0e760bbdde {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.369701] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 871.369701] env[61824]: value = "task-1275691" [ 871.369701] env[61824]: _type = "Task" [ 871.369701] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.383383] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275691, 'name': Destroy_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.384235] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Releasing lock "refresh_cache-b1a770f9-71ea-4f80-a562-c08a313b5753" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.384555] env[61824]: DEBUG nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Instance network_info: |[{"id": "1a99ec03-ba39-4515-827d-6c5c7781f603", "address": "fa:16:3e:3d:dd:da", "network": {"id": "a9d5808f-a48e-482b-8b48-0de699d1ce80", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1537303780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b302783679441b7bc76b291562031d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a99ec03-ba", "ovs_interfaceid": "1a99ec03-ba39-4515-827d-6c5c7781f603", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 871.384850] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] Acquired lock "refresh_cache-b1a770f9-71ea-4f80-a562-c08a313b5753" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.385054] env[61824]: DEBUG nova.network.neutron [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Refreshing network info cache for port 1a99ec03-ba39-4515-827d-6c5c7781f603 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.386372] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:dd:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8abee039-d93e-48a7-8911-6416a3e1ff30', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a99ec03-ba39-4515-827d-6c5c7781f603', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 871.398869] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Creating folder: Project (8b302783679441b7bc76b291562031d7). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 871.404671] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-806c7766-0b28-4e06-9a6f-bdd955e6be42 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.418356] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Created folder: Project (8b302783679441b7bc76b291562031d7) in parent group-v274074. [ 871.418356] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Creating folder: Instances. Parent ref: group-v274174. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 871.418557] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-756ec04a-1be7-4a8d-bb09-feebaac8d20d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.428276] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Created folder: Instances in parent group-v274174. [ 871.428549] env[61824]: DEBUG oslo.service.loopingcall [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.428754] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 871.429310] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8714351c-c077-419d-a231-d51b508d6315 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.456923] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 871.456923] env[61824]: value = "task-1275694" [ 871.456923] env[61824]: _type = "Task" [ 871.456923] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.469713] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275694, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.484154] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275689, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.503187] env[61824]: DEBUG oslo_vmware.api [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142874} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.503511] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.504132] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.504132] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.504132] env[61824]: INFO nova.compute.manager [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 871.504567] env[61824]: DEBUG oslo.service.loopingcall [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.504567] env[61824]: DEBUG nova.compute.manager [-] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 871.504646] env[61824]: DEBUG nova.network.neutron [-] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 871.673216] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b2cde6a-f43c-4418-8a3b-90876244e99d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.681041] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc6f9ce-f985-4ad0-bd2c-49142c926d2a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.728591] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67410351-e920-4800-94c8-450f41871de2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.736679] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0076c033-fc9e-4392-900d-2e517dbc2ba7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.752210] env[61824]: DEBUG nova.compute.provider_tree [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.886603] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275691, 'name': Destroy_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.891130] env[61824]: DEBUG nova.network.neutron [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updated VIF entry in instance network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 871.891567] env[61824]: DEBUG nova.network.neutron [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.948300] env[61824]: DEBUG nova.compute.manager [req-3bfc139b-adc6-462b-88e0-8dd51bf89cc6 req-5dd53584-1657-4dba-8e9d-3c6f554675e5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Received event network-vif-deleted-f28bdbaf-7121-4f64-916f-d52245e55d25 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.948637] env[61824]: INFO nova.compute.manager [req-3bfc139b-adc6-462b-88e0-8dd51bf89cc6 req-5dd53584-1657-4dba-8e9d-3c6f554675e5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Neutron deleted interface f28bdbaf-7121-4f64-916f-d52245e55d25; detaching it from the instance and deleting it from the info cache [ 871.948737] env[61824]: DEBUG nova.network.neutron [req-3bfc139b-adc6-462b-88e0-8dd51bf89cc6 req-5dd53584-1657-4dba-8e9d-3c6f554675e5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.967277] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275694, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.981154] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275689, 'name': CreateVM_Task, 'duration_secs': 0.598883} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.981362] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.982020] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.982282] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.982571] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.982826] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cfb2499-e561-40d2-9ed0-6cb90051fda8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.988503] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 871.988503] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c12731-3bb2-1aea-25c7-966262def155" [ 871.988503] env[61824]: _type = "Task" [ 871.988503] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.996215] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c12731-3bb2-1aea-25c7-966262def155, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.044366] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.044620] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.044778] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.044959] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.045124] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.045541] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.045541] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.045638] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.045794] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.045961] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.046185] env[61824]: DEBUG nova.virt.hardware [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.047050] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb2c669-3364-4494-bd04-a8bbdd0498c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.054986] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb1d712-e679-4c61-bb51-7d0f531d2452 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.068171] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:52:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6076d24d-3c8e-4bbb-ba96-a08fb27a73cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51fba888-670e-4f85-97a5-0133e0624342', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.075741] env[61824]: DEBUG oslo.service.loopingcall [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.076127] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.076230] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-068e6ae5-7274-4acf-a64d-bb882939334a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.097489] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.097489] env[61824]: value = "task-1275695" [ 872.097489] env[61824]: _type = "Task" [ 872.097489] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.107195] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275695, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.248451] env[61824]: DEBUG nova.network.neutron [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Updated VIF entry in instance network info cache for port 1a99ec03-ba39-4515-827d-6c5c7781f603. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.248812] env[61824]: DEBUG nova.network.neutron [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Updating instance_info_cache with network_info: [{"id": "1a99ec03-ba39-4515-827d-6c5c7781f603", "address": "fa:16:3e:3d:dd:da", "network": {"id": "a9d5808f-a48e-482b-8b48-0de699d1ce80", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1537303780-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b302783679441b7bc76b291562031d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8abee039-d93e-48a7-8911-6416a3e1ff30", "external-id": "nsx-vlan-transportzone-654", "segmentation_id": 654, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a99ec03-ba", "ovs_interfaceid": "1a99ec03-ba39-4515-827d-6c5c7781f603", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.255325] env[61824]: DEBUG nova.scheduler.client.report [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.382832] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275691, 'name': Destroy_Task, 'duration_secs': 0.579429} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.383281] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Destroyed the VM [ 872.383667] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Deleting Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 872.384044] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-af62c2f9-e69d-4126-81d5-14edea2de952 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.392197] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 872.392197] env[61824]: value = "task-1275696" [ 872.392197] env[61824]: _type = "Task" [ 872.392197] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.397469] env[61824]: DEBUG oslo_concurrency.lockutils [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.398103] env[61824]: DEBUG nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Received event network-vif-deleted-d4e2c604-4e39-49cc-89a3-b4eed4e7d7ce {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.398710] env[61824]: DEBUG nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Received event network-vif-deleted-cca34907-2b24-40e6-8cab-24bacb75db19 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.398710] env[61824]: INFO nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Neutron deleted interface cca34907-2b24-40e6-8cab-24bacb75db19; detaching it from the instance and deleting it from the info cache [ 872.398710] env[61824]: DEBUG nova.network.neutron [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.409060] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275696, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.413671] env[61824]: DEBUG nova.network.neutron [-] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.439867] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.439867] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.451898] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd93a8c5-068a-41fd-b688-570db9ccb1ee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.463552] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b404e98e-b8f8-455d-a1f0-4f281f7949ad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.480205] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275694, 'name': CreateVM_Task, 'duration_secs': 0.519898} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.480394] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.481064] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.501311] env[61824]: DEBUG nova.compute.manager [req-3bfc139b-adc6-462b-88e0-8dd51bf89cc6 req-5dd53584-1657-4dba-8e9d-3c6f554675e5 service nova] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Detach interface failed, port_id=f28bdbaf-7121-4f64-916f-d52245e55d25, reason: Instance 5806faad-2bab-4fda-a131-566af50c4d7e could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 872.510659] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c12731-3bb2-1aea-25c7-966262def155, 'name': SearchDatastore_Task, 'duration_secs': 0.008784} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.511875] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.511875] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.511875] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.511875] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.512098] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 872.512633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.512633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.512941] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fd26178-6e73-496d-a068-dd5babe38498 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.515401] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-602546f6-8dfd-4658-a7df-3bdd54995cc3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.520951] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 872.520951] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e2d820-0719-9115-c425-ae19493e755e" [ 872.520951] env[61824]: _type = "Task" [ 872.520951] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.525799] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 872.525799] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 872.526828] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5426640f-860b-4a90-b2cb-28bb80c96e20 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.532877] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e2d820-0719-9115-c425-ae19493e755e, 'name': SearchDatastore_Task, 'duration_secs': 0.009178} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.533904] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.534097] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 872.535477] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.535935] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 872.535935] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c0ce09-1060-3fc0-0c2a-93273caf69ee" [ 872.535935] env[61824]: _type = "Task" [ 872.535935] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.543796] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c0ce09-1060-3fc0-0c2a-93273caf69ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.612277] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275695, 'name': CreateVM_Task, 'duration_secs': 0.335797} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.612277] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.612277] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.612277] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.612277] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.612277] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15c0eeb5-864e-4d3d-81ae-53ce63d832e7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.616536] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 872.616536] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a6bcc3-5338-b141-b495-ef814cf5a5b4" [ 872.616536] env[61824]: _type = "Task" [ 872.616536] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.625745] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a6bcc3-5338-b141-b495-ef814cf5a5b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.751984] env[61824]: DEBUG oslo_concurrency.lockutils [req-4f32d389-6862-4f89-aeb4-cb0d202d212b req-f0d48260-894c-4184-8426-4e314a8c8acc service nova] Releasing lock "refresh_cache-b1a770f9-71ea-4f80-a562-c08a313b5753" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.760569] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.761109] env[61824]: DEBUG nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.763726] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.525s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.763914] env[61824]: DEBUG nova.objects.instance [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 872.905318] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275696, 'name': RemoveSnapshot_Task, 'duration_secs': 0.31723} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.905611] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-73fc949b-00f7-4178-87cd-ba9659bb887d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.908112] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Deleted Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 872.908416] env[61824]: DEBUG nova.compute.manager [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.909253] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7e2416-fae3-478c-a8bd-a7c50d3ecb61 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.920668] env[61824]: INFO nova.compute.manager [-] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Took 1.41 seconds to deallocate network for instance. [ 872.926404] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a779eed-2cda-42ae-acfe-922270b517bf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.946441] env[61824]: DEBUG nova.compute.utils [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.969698] env[61824]: DEBUG nova.compute.manager [req-67f6eb0f-6d9f-42d7-b93d-5c2926af644d req-63d50a19-93fd-4239-bde2-f59d7768262f service nova] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Detach interface failed, port_id=cca34907-2b24-40e6-8cab-24bacb75db19, reason: Instance cabad7a6-206f-4e02-be3c-e18654396b7e could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 873.047528] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c0ce09-1060-3fc0-0c2a-93273caf69ee, 'name': SearchDatastore_Task, 'duration_secs': 0.008147} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.048450] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c63168b2-7102-4b04-90a1-5989d38b55f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.054606] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 873.054606] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52357373-c213-ab70-f78a-7ec705fe74a6" [ 873.054606] env[61824]: _type = "Task" [ 873.054606] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.062431] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52357373-c213-ab70-f78a-7ec705fe74a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.126602] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a6bcc3-5338-b141-b495-ef814cf5a5b4, 'name': SearchDatastore_Task, 'duration_secs': 0.009107} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.126962] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.127262] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.127535] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.127799] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.127964] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.128329] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dee553b-1201-4b9a-a3e0-e613d2047c4d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.136315] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.136471] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.137184] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36ed0de7-025b-44de-8bd4-81d97b831302 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.142361] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 873.142361] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5236c816-f8d7-204e-5ff0-f309584fc17a" [ 873.142361] env[61824]: _type = "Task" [ 873.142361] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.150245] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5236c816-f8d7-204e-5ff0-f309584fc17a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.270019] env[61824]: DEBUG nova.compute.utils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.270019] env[61824]: DEBUG nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.270019] env[61824]: DEBUG nova.network.neutron [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 873.299485] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "interface-5569eb0a-fd42-4e2d-b805-bb40e3264a65-a608c7e2-51a6-4791-b00b-570fb8a353b0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.300267] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-5569eb0a-fd42-4e2d-b805-bb40e3264a65-a608c7e2-51a6-4791-b00b-570fb8a353b0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.300865] env[61824]: DEBUG nova.objects.instance [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lazy-loading 'flavor' on Instance uuid 5569eb0a-fd42-4e2d-b805-bb40e3264a65 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.337728] env[61824]: DEBUG nova.policy [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6a2f1ffe8934519acdcaca77e2a55d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e51dd2f578344aa8afa5d6beccb54f1e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.422513] env[61824]: INFO nova.compute.manager [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Shelve offloading [ 873.425717] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.426072] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82ffa123-15be-47de-b16c-13014255fd1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.432956] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 873.432956] env[61824]: value = "task-1275697" [ 873.432956] env[61824]: _type = "Task" [ 873.432956] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.442466] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 873.442649] env[61824]: DEBUG nova.compute.manager [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.443438] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e7571a-9496-40e0-82f3-5085bfd80b93 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.450415] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.451295] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.451606] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.451785] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.451962] env[61824]: DEBUG nova.network.neutron [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 873.571593] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52357373-c213-ab70-f78a-7ec705fe74a6, 'name': SearchDatastore_Task, 'duration_secs': 0.008443} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.574886] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.574886] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] c6552176-1128-473b-ac4a-f3b55428fd9e/c6552176-1128-473b-ac4a-f3b55428fd9e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 873.574886] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.574886] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.575117] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-782c1749-9f56-4a94-b1bd-c2a790b8ad6c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.578192] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-92405d69-6ed5-4b1d-92c8-c45d968d437c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.588677] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 873.588677] env[61824]: value = "task-1275698" [ 873.588677] env[61824]: _type = "Task" [ 873.588677] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.591260] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.591586] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.598600] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ebbc625-07aa-4001-9b33-acfb6d6be79a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.617021] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.617021] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 873.617021] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52ab3c1e-55aa-7784-2d66-a0ea04d44393" [ 873.617021] env[61824]: _type = "Task" [ 873.617021] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.625257] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ab3c1e-55aa-7784-2d66-a0ea04d44393, 'name': SearchDatastore_Task, 'duration_secs': 0.008369} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.626340] env[61824]: DEBUG nova.network.neutron [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Successfully created port: 542a267a-6d8a-4233-8c5a-64d5b20c7f72 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.629995] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3845f79e-92a9-45bf-b3f3-0e7bf1dbcc27 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.636260] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 873.636260] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]529b3bf3-232b-b959-fa89-865484c1b189" [ 873.636260] env[61824]: _type = "Task" [ 873.636260] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.644798] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529b3bf3-232b-b959-fa89-865484c1b189, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.656125] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5236c816-f8d7-204e-5ff0-f309584fc17a, 'name': SearchDatastore_Task, 'duration_secs': 0.007776} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.657085] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8134d08c-f0b2-477a-85f9-c9c2fda13cc7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.662859] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 873.662859] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525551f9-6fdc-7714-5c31-d998965ba67e" [ 873.662859] env[61824]: _type = "Task" [ 873.662859] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.672990] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525551f9-6fdc-7714-5c31-d998965ba67e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.777289] env[61824]: DEBUG nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.779773] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b81a923d-bca1-48d5-bf42-e068061101b6 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.780873] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.528s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.781053] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.783602] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.753s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.783913] env[61824]: DEBUG nova.objects.instance [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lazy-loading 'resources' on Instance uuid d9bb3172-c058-4e59-a719-d8aa340fba71 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.817449] env[61824]: INFO nova.scheduler.client.report [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted allocations for instance a33e2573-6d2f-4d07-89d4-f1c6d9d9b382 [ 874.011677] env[61824]: DEBUG nova.objects.instance [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lazy-loading 'pci_requests' on Instance uuid 5569eb0a-fd42-4e2d-b805-bb40e3264a65 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.101857] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44425} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.102313] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] c6552176-1128-473b-ac4a-f3b55428fd9e/c6552176-1128-473b-ac4a-f3b55428fd9e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 874.102546] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 874.102820] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-557fa2b8-7e43-404f-a44f-caddc335064b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.109623] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 874.109623] env[61824]: value = "task-1275699" [ 874.109623] env[61824]: _type = "Task" [ 874.109623] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.117655] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275699, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.147463] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529b3bf3-232b-b959-fa89-865484c1b189, 'name': SearchDatastore_Task, 'duration_secs': 0.008551} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.147732] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.148039] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] b1a770f9-71ea-4f80-a562-c08a313b5753/b1a770f9-71ea-4f80-a562-c08a313b5753.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.148283] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb6c416a-082b-4c39-b507-ac12ff11c478 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.155159] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 874.155159] env[61824]: value = "task-1275700" [ 874.155159] env[61824]: _type = "Task" [ 874.155159] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.164258] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275700, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.173060] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525551f9-6fdc-7714-5c31-d998965ba67e, 'name': SearchDatastore_Task, 'duration_secs': 0.009304} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.173370] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.173604] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.174258] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb1660f7-6b25-4e24-a047-c05006e1d3f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.180646] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 874.180646] env[61824]: value = "task-1275701" [ 874.180646] env[61824]: _type = "Task" [ 874.180646] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.188788] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.330419] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c1c4d0cf-79a9-4a28-aa00-7c7f7eb06097 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "a33e2573-6d2f-4d07-89d4-f1c6d9d9b382" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.573s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.485481] env[61824]: DEBUG nova.network.neutron [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.514782] env[61824]: DEBUG nova.objects.base [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Object Instance<5569eb0a-fd42-4e2d-b805-bb40e3264a65> lazy-loaded attributes: flavor,pci_requests {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 874.515049] env[61824]: DEBUG nova.network.neutron [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 874.544835] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.545126] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.545374] env[61824]: INFO nova.compute.manager [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Attaching volume fd8092a8-d5fb-41a2-b612-00a7cff5fbe7 to /dev/sdc [ 874.586753] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d06bb8-872f-476e-a488-6b650d992dd5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.598371] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70a78a3-8a31-4244-87af-701e31f7d6a5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.610717] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed80e8f9-f004-4808-bdb2-cc5c59af2afc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.614914] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617eb4a6-dbd4-424f-8b49-8141a25495ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.626637] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275699, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064329} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.654750] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.659655] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ba19e4-ee58-4530-a5d6-69e8405fd45b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.662943] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d216bd4-5eb3-4e2f-bbda-107e1f38e16f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.666091] env[61824]: DEBUG nova.virt.block_device [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updating existing volume attachment record: 8793006f-3a35-4fbd-a822-f8b893c3848c {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 874.690580] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f3739c-26af-4a81-a374-de0e725f83d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.694871] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275700, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.704036] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] c6552176-1128-473b-ac4a-f3b55428fd9e/c6552176-1128-473b-ac4a-f3b55428fd9e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.707901] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b81d8f8-e698-4b2e-8072-a4b31274288e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.727165] env[61824]: DEBUG nova.policy [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '316c089785a44aa897040f3c0ef7a071', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0924bfa41b5e42d68818557d2b0ace7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.739650] env[61824]: DEBUG nova.compute.provider_tree [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.745751] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 874.745751] env[61824]: value = "task-1275702" [ 874.745751] env[61824]: _type = "Task" [ 874.745751] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.747017] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275701, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.757891] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275702, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.794055] env[61824]: DEBUG nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.825234] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.825474] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.825632] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.825815] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.825988] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.826241] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.826352] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.826525] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.826685] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.826859] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.827044] env[61824]: DEBUG nova.virt.hardware [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.827990] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3457c688-40f0-4d86-8173-4943390ea5cf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.836285] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cef7839-3e83-4e68-bbaa-e70b97967e28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.988953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.175386] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275700, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629699} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.175677] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] b1a770f9-71ea-4f80-a562-c08a313b5753/b1a770f9-71ea-4f80-a562-c08a313b5753.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.175970] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.176191] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-372fbeb1-d7ee-44ea-8b53-2e7b7bed25a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.184067] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 875.184067] env[61824]: value = "task-1275704" [ 875.184067] env[61824]: _type = "Task" [ 875.184067] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.198670] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.208328] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.633239} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.208479] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.208613] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.208881] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-48b5797f-ec1e-4f8c-bf86-c7fa7e7986c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.216200] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 875.216200] env[61824]: value = "task-1275705" [ 875.216200] env[61824]: _type = "Task" [ 875.216200] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.228264] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275705, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.248201] env[61824]: DEBUG nova.scheduler.client.report [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.262444] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275702, 'name': ReconfigVM_Task, 'duration_secs': 0.303766} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.263062] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Reconfigured VM instance instance-0000004e to attach disk [datastore1] c6552176-1128-473b-ac4a-f3b55428fd9e/c6552176-1128-473b-ac4a-f3b55428fd9e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.263845] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e1054cb-1490-4d9e-aec5-d8bc43846692 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.271185] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 875.271185] env[61824]: value = "task-1275706" [ 875.271185] env[61824]: _type = "Task" [ 875.271185] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.282415] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275706, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.364215] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.365588] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2333d19-e299-43ef-bf48-94199e348140 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.374031] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.374464] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-65fb0288-33b8-4de2-b332-335541cdccb9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.451105] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 875.451105] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 875.451105] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleting the datastore file [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.451105] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-543a04bc-dd47-493d-a52c-98dbc5146366 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.459080] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 875.459080] env[61824]: value = "task-1275708" [ 875.459080] env[61824]: _type = "Task" [ 875.459080] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.467332] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275708, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.611369] env[61824]: DEBUG nova.network.neutron [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Successfully updated port: 542a267a-6d8a-4233-8c5a-64d5b20c7f72 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.675154] env[61824]: DEBUG nova.compute.manager [req-1514fe18-00d2-4aac-8df3-59f8544c9856 req-27536833-a013-47bf-a67c-edead0546916 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-vif-unplugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.675154] env[61824]: DEBUG oslo_concurrency.lockutils [req-1514fe18-00d2-4aac-8df3-59f8544c9856 req-27536833-a013-47bf-a67c-edead0546916 service nova] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.675154] env[61824]: DEBUG oslo_concurrency.lockutils [req-1514fe18-00d2-4aac-8df3-59f8544c9856 req-27536833-a013-47bf-a67c-edead0546916 service nova] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.675154] env[61824]: DEBUG oslo_concurrency.lockutils [req-1514fe18-00d2-4aac-8df3-59f8544c9856 req-27536833-a013-47bf-a67c-edead0546916 service nova] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.675154] env[61824]: DEBUG nova.compute.manager [req-1514fe18-00d2-4aac-8df3-59f8544c9856 req-27536833-a013-47bf-a67c-edead0546916 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] No waiting events found dispatching network-vif-unplugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.675481] env[61824]: WARNING nova.compute.manager [req-1514fe18-00d2-4aac-8df3-59f8544c9856 req-27536833-a013-47bf-a67c-edead0546916 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received unexpected event network-vif-unplugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b for instance with vm_state shelved and task_state shelving_offloading. [ 875.693849] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.217323} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.694167] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.695041] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74a8f9c-60d9-4cdd-a5f2-f72d19d1cd7f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.720090] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] b1a770f9-71ea-4f80-a562-c08a313b5753/b1a770f9-71ea-4f80-a562-c08a313b5753.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.720485] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-223609b9-3829-46b1-9af2-d9b05b4dfcf5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.747860] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275705, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063687} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.749187] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.749577] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 875.749577] env[61824]: value = "task-1275709" [ 875.749577] env[61824]: _type = "Task" [ 875.749577] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.750302] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36cb65b6-1299-4b37-a8c6-812491615142 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.757254] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.974s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.772648] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.279s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.772941] env[61824]: DEBUG nova.objects.instance [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'resources' on Instance uuid cabad7a6-206f-4e02-be3c-e18654396b7e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.787286] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.800168] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f6ba1fd-aacf-4edf-891d-baba12746048 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.815210] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.817660] env[61824]: INFO nova.scheduler.client.report [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Deleted allocations for instance d9bb3172-c058-4e59-a719-d8aa340fba71 [ 875.819855] env[61824]: DEBUG nova.compute.manager [req-873d0d55-adc2-4611-a985-467b42327d5a req-7617ffae-db95-4f73-bc7c-c362ccafceeb service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Received event network-vif-plugged-542a267a-6d8a-4233-8c5a-64d5b20c7f72 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.820081] env[61824]: DEBUG oslo_concurrency.lockutils [req-873d0d55-adc2-4611-a985-467b42327d5a req-7617ffae-db95-4f73-bc7c-c362ccafceeb service nova] Acquiring lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.820291] env[61824]: DEBUG oslo_concurrency.lockutils [req-873d0d55-adc2-4611-a985-467b42327d5a req-7617ffae-db95-4f73-bc7c-c362ccafceeb service nova] Lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.820710] env[61824]: DEBUG oslo_concurrency.lockutils [req-873d0d55-adc2-4611-a985-467b42327d5a req-7617ffae-db95-4f73-bc7c-c362ccafceeb service nova] Lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.820710] env[61824]: DEBUG nova.compute.manager [req-873d0d55-adc2-4611-a985-467b42327d5a req-7617ffae-db95-4f73-bc7c-c362ccafceeb service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] No waiting events found dispatching network-vif-plugged-542a267a-6d8a-4233-8c5a-64d5b20c7f72 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.820775] env[61824]: WARNING nova.compute.manager [req-873d0d55-adc2-4611-a985-467b42327d5a req-7617ffae-db95-4f73-bc7c-c362ccafceeb service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Received unexpected event network-vif-plugged-542a267a-6d8a-4233-8c5a-64d5b20c7f72 for instance with vm_state building and task_state spawning. [ 875.830453] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275706, 'name': Rename_Task, 'duration_secs': 0.157244} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.832083] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.832359] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 875.832359] env[61824]: value = "task-1275710" [ 875.832359] env[61824]: _type = "Task" [ 875.832359] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.833134] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45285c06-5354-4923-9d5e-f50761d17c49 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.845430] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 875.845430] env[61824]: value = "task-1275711" [ 875.845430] env[61824]: _type = "Task" [ 875.845430] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.850036] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.859512] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.969324] env[61824]: DEBUG oslo_vmware.api [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275708, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14112} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.969634] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.969823] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.970668] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.989768] env[61824]: INFO nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted allocations for instance 7be79422-c624-4b4a-884f-29b13b9d81d2 [ 876.085305] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e28297-7cce-4dce-855c-e1f11b035499 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.093414] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a48840e-eddf-4040-864d-ea05c4cb7aaf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.125741] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.125895] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.126064] env[61824]: DEBUG nova.network.neutron [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.127935] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5c1fa4-e0a7-46af-a898-2f3e329cbeeb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.136988] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91510620-9590-45cc-b40d-0aabd41af1d2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.151855] env[61824]: DEBUG nova.compute.provider_tree [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.270419] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275709, 'name': ReconfigVM_Task, 'duration_secs': 0.489436} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.270419] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Reconfigured VM instance instance-0000004f to attach disk [datastore1] b1a770f9-71ea-4f80-a562-c08a313b5753/b1a770f9-71ea-4f80-a562-c08a313b5753.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.270419] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fae45d97-0f96-4a40-8d47-7f341f5c8ed0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.277025] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 876.277025] env[61824]: value = "task-1275712" [ 876.277025] env[61824]: _type = "Task" [ 876.277025] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.286575] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275712, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.328810] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e02f002e-c77e-4bd1-a04a-f819ffef293b tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "d9bb3172-c058-4e59-a719-d8aa340fba71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.764s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.344291] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.358734] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275711, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.494630] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.654793] env[61824]: DEBUG nova.scheduler.client.report [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.790758] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275712, 'name': Rename_Task, 'duration_secs': 0.142246} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.791075] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.793504] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0910afe8-44be-48ef-a956-5bdc00cf3c9f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.800863] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 876.800863] env[61824]: value = "task-1275713" [ 876.800863] env[61824]: _type = "Task" [ 876.800863] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.810365] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.845644] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275710, 'name': ReconfigVM_Task, 'duration_secs': 0.750628} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.846010] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558/7e73f8b5-a138-4455-a392-9a2b2b860558.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.846749] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-059c27ab-3876-4214-b969-fd12f9be82b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.856601] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 876.856601] env[61824]: value = "task-1275714" [ 876.856601] env[61824]: _type = "Task" [ 876.856601] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.867712] env[61824]: DEBUG oslo_vmware.api [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275711, 'name': PowerOnVM_Task, 'duration_secs': 0.831735} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.868433] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 876.868946] env[61824]: INFO nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Took 10.93 seconds to spawn the instance on the hypervisor. [ 876.868946] env[61824]: DEBUG nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 876.869663] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aefc59c-8636-4ec8-984c-62ca197c51b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.876252] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275714, 'name': Rename_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.898823] env[61824]: DEBUG nova.network.neutron [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.163185] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.390s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.167829] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.717s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.168253] env[61824]: DEBUG nova.objects.instance [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lazy-loading 'resources' on Instance uuid 5806faad-2bab-4fda-a131-566af50c4d7e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.213229] env[61824]: DEBUG nova.network.neutron [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance_info_cache with network_info: [{"id": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "address": "fa:16:3e:ad:9e:74", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap542a267a-6d", "ovs_interfaceid": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.217557] env[61824]: INFO nova.scheduler.client.report [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted allocations for instance cabad7a6-206f-4e02-be3c-e18654396b7e [ 877.258245] env[61824]: DEBUG nova.network.neutron [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Successfully updated port: a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.314662] env[61824]: DEBUG oslo_vmware.api [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275713, 'name': PowerOnVM_Task, 'duration_secs': 0.479391} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.315425] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.315697] env[61824]: INFO nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Took 8.87 seconds to spawn the instance on the hypervisor. [ 877.316170] env[61824]: DEBUG nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 877.316845] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c5e92d-c08a-4988-ada5-36ba49d10712 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.368425] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275714, 'name': Rename_Task, 'duration_secs': 0.1963} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.368822] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.369098] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ddb620f6-cd80-4017-9f29-dd3298d05bc2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.376102] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 877.376102] env[61824]: value = "task-1275716" [ 877.376102] env[61824]: _type = "Task" [ 877.376102] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.383601] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275716, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.393166] env[61824]: INFO nova.compute.manager [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Took 25.65 seconds to build instance. [ 877.719892] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.720240] env[61824]: DEBUG nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Instance network_info: |[{"id": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "address": "fa:16:3e:ad:9e:74", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap542a267a-6d", "ovs_interfaceid": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.720763] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:9e:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd646f9d5-d2ad-4c22-bea5-85a965334de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '542a267a-6d8a-4233-8c5a-64d5b20c7f72', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.729340] env[61824]: DEBUG oslo.service.loopingcall [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.733964] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 877.734596] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02575969-7338-43f6-ad14-de359766850f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "cabad7a6-206f-4e02-be3c-e18654396b7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.215s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.738047] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b39f33c6-1b69-4e5a-be1b-2534f45d3465 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.762055] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.762235] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.763027] env[61824]: DEBUG nova.network.neutron [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.770514] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.770514] env[61824]: value = "task-1275717" [ 877.770514] env[61824]: _type = "Task" [ 877.770514] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.787570] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275717, 'name': CreateVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.839801] env[61824]: INFO nova.compute.manager [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Took 23.67 seconds to build instance. [ 877.888057] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275716, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.896260] env[61824]: DEBUG oslo_concurrency.lockutils [None req-20a5b994-7483-4289-a939-9d9c1d990e6e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.167s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.903576] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "9f031313-766b-4696-9bc6-75ced078bfde" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.903824] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "9f031313-766b-4696-9bc6-75ced078bfde" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.935912] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b37ae2-4233-4ca1-b5e0-af789921a71b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.945175] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22ce43b-c123-4773-85e5-cf848ebccadd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.979217] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0036b2e-4401-4b1d-83ba-f5ae3664abdf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.987541] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b8bf0c-a40e-46c7-9080-1eb3c4c9f65e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.005321] env[61824]: DEBUG nova.compute.provider_tree [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.148257] env[61824]: DEBUG nova.compute.manager [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.148479] env[61824]: DEBUG nova.compute.manager [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing instance network info cache due to event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.148698] env[61824]: DEBUG oslo_concurrency.lockutils [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.148843] env[61824]: DEBUG oslo_concurrency.lockutils [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.149198] env[61824]: DEBUG nova.network.neutron [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.267314] env[61824]: DEBUG nova.compute.manager [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Received event network-changed-542a267a-6d8a-4233-8c5a-64d5b20c7f72 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.267314] env[61824]: DEBUG nova.compute.manager [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Refreshing instance network info cache due to event network-changed-542a267a-6d8a-4233-8c5a-64d5b20c7f72. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.267314] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Acquiring lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.267314] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Acquired lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.267314] env[61824]: DEBUG nova.network.neutron [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Refreshing network info cache for port 542a267a-6d8a-4233-8c5a-64d5b20c7f72 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.287439] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275717, 'name': CreateVM_Task, 'duration_secs': 0.317315} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.287534] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.288211] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.288370] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.288686] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.288948] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-918f9504-31ff-43fb-9d16-9e7cf2816b44 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.293903] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 878.293903] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5270f80c-e003-f2e7-bf6e-c18b7d4d631e" [ 878.293903] env[61824]: _type = "Task" [ 878.293903] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.303220] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5270f80c-e003-f2e7-bf6e-c18b7d4d631e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.316101] env[61824]: WARNING nova.network.neutron [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] 34841295-0562-4caa-8275-c3695a63d719 already exists in list: networks containing: ['34841295-0562-4caa-8275-c3695a63d719']. ignoring it [ 878.341983] env[61824]: DEBUG oslo_concurrency.lockutils [None req-263fa365-9091-45cb-a958-71e5094d26d0 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.177s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.390330] env[61824]: DEBUG oslo_vmware.api [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275716, 'name': PowerOnVM_Task, 'duration_secs': 0.729269} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.391246] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.391246] env[61824]: DEBUG nova.compute.manager [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.391904] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d66a3329-4216-49e6-905e-7a880a41fb60 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.406092] env[61824]: DEBUG nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 879.202721] env[61824]: DEBUG nova.scheduler.client.report [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.208846] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "c6552176-1128-473b-ac4a-f3b55428fd9e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.209205] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.209346] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "c6552176-1128-473b-ac4a-f3b55428fd9e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.209495] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.209735] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.211541] env[61824]: INFO nova.compute.manager [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Terminating instance [ 879.214909] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "b1a770f9-71ea-4f80-a562-c08a313b5753" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.215120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.215314] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "b1a770f9-71ea-4f80-a562-c08a313b5753-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.215496] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.215680] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.217391] env[61824]: INFO nova.compute.manager [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Terminating instance [ 879.219408] env[61824]: DEBUG nova.compute.manager [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.220038] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 879.230215] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 879.230215] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274178', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'name': 'volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'serial': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 879.231171] env[61824]: DEBUG nova.compute.manager [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 879.232419] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 879.237280] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93d0909d-eaa8-46cc-a323-07eaad5d1bca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.240722] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1a088e-545e-4425-aad4-fcdf5ae27b28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.244511] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502c5ca5-2736-4ceb-b86a-8e58ff3eaf9f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.258364] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5270f80c-e003-f2e7-bf6e-c18b7d4d631e, 'name': SearchDatastore_Task, 'duration_secs': 0.01036} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.273704] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.274390] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.274390] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.274390] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.274801] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.275079] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.275896] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 879.276880] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.277131] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-098f1595-822a-4942-9f1b-5d0eb8071099 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.279048] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-acba0617-03ad-45bd-8759-8c9d0541026c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.281217] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71abca7-ff67-4874-a1f2-67c1f5451592 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.284436] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c392536-7f75-4cb7-bb64-d0ff7cc2f3d9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.290878] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.319607] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7/volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.322967] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76907c32-bce5-4a49-8d35-ef7762b6a190 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.338090] env[61824]: DEBUG oslo_vmware.api [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 879.338090] env[61824]: value = "task-1275718" [ 879.338090] env[61824]: _type = "Task" [ 879.338090] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.338090] env[61824]: DEBUG oslo_vmware.api [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 879.338090] env[61824]: value = "task-1275719" [ 879.338090] env[61824]: _type = "Task" [ 879.338090] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.338090] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.338090] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.344596] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d531cde-2fa1-4332-bd10-db802137cbc4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.364592] env[61824]: DEBUG oslo_vmware.api [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 879.364592] env[61824]: value = "task-1275720" [ 879.364592] env[61824]: _type = "Task" [ 879.364592] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.377867] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 879.377867] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a2b15d-9cef-9fa3-6bb3-993fdd5b6092" [ 879.377867] env[61824]: _type = "Task" [ 879.377867] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.378335] env[61824]: DEBUG oslo_vmware.api [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.378557] env[61824]: DEBUG oslo_vmware.api [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.387901] env[61824]: DEBUG oslo_vmware.api [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275720, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.393183] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a2b15d-9cef-9fa3-6bb3-993fdd5b6092, 'name': SearchDatastore_Task, 'duration_secs': 0.010004} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.393924] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f4a208f-b371-43f6-bdfe-0979dbfff6e0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.399648] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 879.399648] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52bfed5b-a584-ce10-7c2d-5cbf0163390f" [ 879.399648] env[61824]: _type = "Task" [ 879.399648] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.411945] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bfed5b-a584-ce10-7c2d-5cbf0163390f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.514494] env[61824]: DEBUG nova.network.neutron [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updated VIF entry in instance network info cache for port 542a267a-6d8a-4233-8c5a-64d5b20c7f72. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.514865] env[61824]: DEBUG nova.network.neutron [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance_info_cache with network_info: [{"id": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "address": "fa:16:3e:ad:9e:74", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap542a267a-6d", "ovs_interfaceid": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.708420] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.540s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.711082] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.217s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.711372] env[61824]: DEBUG nova.objects.instance [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'resources' on Instance uuid 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.713374] env[61824]: DEBUG nova.network.neutron [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "address": "fa:16:3e:da:35:5b", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa608c7e2-51", "ovs_interfaceid": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.728571] env[61824]: INFO nova.scheduler.client.report [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted allocations for instance 5806faad-2bab-4fda-a131-566af50c4d7e [ 879.789520] env[61824]: DEBUG nova.network.neutron [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updated VIF entry in instance network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.789880] env[61824]: DEBUG nova.network.neutron [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.852365] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.858761] env[61824]: DEBUG oslo_vmware.api [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275718, 'name': PowerOffVM_Task, 'duration_secs': 0.255884} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.861738] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.862300] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 879.862300] env[61824]: DEBUG oslo_vmware.api [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275719, 'name': PowerOffVM_Task, 'duration_secs': 0.194147} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.862443] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2aa028a-8972-4177-875e-e58e6e26d2bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.863953] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.864122] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 879.864350] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c598841d-1a4f-40bb-a95d-bfcaf633bbb8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.873597] env[61824]: DEBUG oslo_vmware.api [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275720, 'name': ReconfigVM_Task, 'duration_secs': 0.358807} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.873843] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfigured VM instance instance-00000042 to attach disk [datastore2] volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7/volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.878552] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e567c39-c9a9-4909-bdb4-ae50074bae65 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.894549] env[61824]: DEBUG oslo_vmware.api [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 879.894549] env[61824]: value = "task-1275723" [ 879.894549] env[61824]: _type = "Task" [ 879.894549] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.902537] env[61824]: DEBUG oslo_vmware.api [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275723, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.910149] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bfed5b-a584-ce10-7c2d-5cbf0163390f, 'name': SearchDatastore_Task, 'duration_secs': 0.00921} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.910434] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.910742] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 09432e72-c10a-4a7a-93af-af0d371fb871/09432e72-c10a-4a7a-93af-af0d371fb871.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 879.910950] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2b989ed-380b-4b4b-bb2d-19e5de8e22b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.917445] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 879.917445] env[61824]: value = "task-1275724" [ 879.917445] env[61824]: _type = "Task" [ 879.917445] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.925418] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275724, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.964162] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.964559] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.964860] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Deleting the datastore file [datastore1] b1a770f9-71ea-4f80-a562-c08a313b5753 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.967363] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e55873ac-25b7-4967-bd2f-b83347237a8b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.970776] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.971482] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.972084] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleting the datastore file [datastore1] c6552176-1128-473b-ac4a-f3b55428fd9e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.972466] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a8b6253-68ac-4ebb-90d9-9ad566c4eb48 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.980064] env[61824]: DEBUG oslo_vmware.api [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for the task: (returnval){ [ 879.980064] env[61824]: value = "task-1275725" [ 879.980064] env[61824]: _type = "Task" [ 879.980064] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.981407] env[61824]: DEBUG oslo_vmware.api [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 879.981407] env[61824]: value = "task-1275726" [ 879.981407] env[61824]: _type = "Task" [ 879.981407] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.992483] env[61824]: DEBUG oslo_vmware.api [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275725, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.995767] env[61824]: DEBUG oslo_vmware.api [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.017661] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Releasing lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.018136] env[61824]: DEBUG nova.compute.manager [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-vif-plugged-a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.018395] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Acquiring lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.018675] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.018877] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.019108] env[61824]: DEBUG nova.compute.manager [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] No waiting events found dispatching network-vif-plugged-a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.019300] env[61824]: WARNING nova.compute.manager [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received unexpected event network-vif-plugged-a608c7e2-51a6-4791-b00b-570fb8a353b0 for instance with vm_state active and task_state None. [ 880.019469] env[61824]: DEBUG nova.compute.manager [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-changed-a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.019628] env[61824]: DEBUG nova.compute.manager [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing instance network info cache due to event network-changed-a608c7e2-51a6-4791-b00b-570fb8a353b0. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.019850] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.216353] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.217094] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.217269] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.217806] env[61824]: DEBUG nova.objects.instance [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'numa_topology' on Instance uuid 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.218937] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.219162] env[61824]: DEBUG nova.network.neutron [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing network info cache for port a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.220752] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b561b29e-459b-4cb3-93ed-00888f074386 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.242194] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.242420] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.243887] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.244184] env[61824]: DEBUG nova.virt.hardware [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.250530] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Reconfiguring VM to attach interface {{(pid=61824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 880.251237] env[61824]: DEBUG oslo_concurrency.lockutils [None req-29a24628-3b23-4e84-8b10-98cc878efaf7 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "5806faad-2bab-4fda-a131-566af50c4d7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.885s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.253046] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b20356a5-88a3-4480-8334-1f1b591b16e7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.273622] env[61824]: DEBUG oslo_vmware.api [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 880.273622] env[61824]: value = "task-1275727" [ 880.273622] env[61824]: _type = "Task" [ 880.273622] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.284191] env[61824]: DEBUG oslo_vmware.api [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275727, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.292745] env[61824]: DEBUG oslo_concurrency.lockutils [req-ad9f4dcf-b42d-416c-a0e7-cd6e8168b9a2 req-18082632-b53d-4002-8e20-24ee67907b5e service nova] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.407331] env[61824]: DEBUG oslo_vmware.api [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275723, 'name': ReconfigVM_Task, 'duration_secs': 0.152938} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.407757] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274178', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'name': 'volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'serial': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 880.431828] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275724, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.474115] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.474421] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.494067] env[61824]: DEBUG oslo_vmware.api [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Task: {'id': task-1275725, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163623} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.496944] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.497240] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.497433] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.497609] env[61824]: INFO nova.compute.manager [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Took 1.28 seconds to destroy the instance on the hypervisor. [ 880.498454] env[61824]: DEBUG oslo.service.loopingcall [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.498674] env[61824]: DEBUG oslo_vmware.api [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199911} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.499111] env[61824]: DEBUG nova.compute.manager [-] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.499209] env[61824]: DEBUG nova.network.neutron [-] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.500804] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.500989] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.501183] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.501355] env[61824]: INFO nova.compute.manager [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Took 1.27 seconds to destroy the instance on the hypervisor. [ 880.501573] env[61824]: DEBUG oslo.service.loopingcall [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.501759] env[61824]: DEBUG nova.compute.manager [-] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 880.501849] env[61824]: DEBUG nova.network.neutron [-] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.724387] env[61824]: DEBUG nova.objects.base [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Object Instance<7be79422-c624-4b4a-884f-29b13b9d81d2> lazy-loaded attributes: resources,numa_topology {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 880.787959] env[61824]: DEBUG oslo_vmware.api [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.891123] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.891988] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.939452] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275724, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541679} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.940118] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 09432e72-c10a-4a7a-93af-af0d371fb871/09432e72-c10a-4a7a-93af-af0d371fb871.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 880.940471] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.945042] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d795a2c0-a972-4765-ae84-564847690e6f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.959312] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 880.959312] env[61824]: value = "task-1275728" [ 880.959312] env[61824]: _type = "Task" [ 880.959312] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.984342] env[61824]: DEBUG nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.991031] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275728, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.025696] env[61824]: DEBUG nova.network.neutron [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updated VIF entry in instance network info cache for port a608c7e2-51a6-4791-b00b-570fb8a353b0. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.026320] env[61824]: DEBUG nova.network.neutron [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "address": "fa:16:3e:da:35:5b", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa608c7e2-51", "ovs_interfaceid": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.075946] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cca7648-9e19-4e0c-a781-8207477f6288 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.084325] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6fff22-95b8-48a5-b5b0-48dff5b19a0b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.118821] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4118e9f5-35b5-4999-aa0b-8d8529452760 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.126928] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe9a042-2a75-49ca-a3e8-9d61d8cb01b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.141143] env[61824]: DEBUG nova.compute.provider_tree [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.285668] env[61824]: DEBUG oslo_vmware.api [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275727, 'name': ReconfigVM_Task, 'duration_secs': 0.913113} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.286133] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.286368] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Reconfigured VM to attach interface {{(pid=61824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 881.348292] env[61824]: DEBUG nova.compute.manager [req-0eb2c3ba-4dca-473c-a902-06a2306a6323 req-839c47e4-ae72-4293-8d1f-a4623849f4c2 service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Received event network-vif-deleted-1a99ec03-ba39-4515-827d-6c5c7781f603 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.348292] env[61824]: INFO nova.compute.manager [req-0eb2c3ba-4dca-473c-a902-06a2306a6323 req-839c47e4-ae72-4293-8d1f-a4623849f4c2 service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Neutron deleted interface 1a99ec03-ba39-4515-827d-6c5c7781f603; detaching it from the instance and deleting it from the info cache [ 881.348292] env[61824]: DEBUG nova.network.neutron [req-0eb2c3ba-4dca-473c-a902-06a2306a6323 req-839c47e4-ae72-4293-8d1f-a4623849f4c2 service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.395773] env[61824]: DEBUG nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 881.468303] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275728, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142517} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.468602] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.469381] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27669a83-cc94-4f4f-ac78-9e3246ae6fa4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.482227] env[61824]: DEBUG nova.network.neutron [-] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.491909] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 09432e72-c10a-4a7a-93af-af0d371fb871/09432e72-c10a-4a7a-93af-af0d371fb871.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.492872] env[61824]: INFO nova.compute.manager [-] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Took 0.99 seconds to deallocate network for instance. [ 881.495031] env[61824]: DEBUG nova.network.neutron [-] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.495874] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a324a1a7-534a-4c83-8d7b-1f75d92fa0d0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.513917] env[61824]: DEBUG nova.objects.instance [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lazy-loading 'flavor' on Instance uuid cbded637-56b9-4048-b66e-1edf6311f435 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.523555] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 881.523555] env[61824]: value = "task-1275729" [ 881.523555] env[61824]: _type = "Task" [ 881.523555] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.527810] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.529943] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed3abe3b-ceb2-4575-913b-f6b61893ebff req-2ad4a900-8048-47e2-9e1a-3df0bd3a7c38 service nova] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.533576] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275729, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.666787] env[61824]: ERROR nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [req-77a53cec-5137-4365-a862-84b088b14de0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8d9ece0e-729c-4d69-b741-d89969200860. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-77a53cec-5137-4365-a862-84b088b14de0"}]} [ 881.689339] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Refreshing inventories for resource provider 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 881.704626] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating ProviderTree inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 881.704626] env[61824]: DEBUG nova.compute.provider_tree [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.714760] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Refreshing aggregate associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, aggregates: None {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 881.733306] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Refreshing trait associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 881.790743] env[61824]: DEBUG oslo_concurrency.lockutils [None req-42712a7f-fe70-4904-b2bc-a330f568b934 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-5569eb0a-fd42-4e2d-b805-bb40e3264a65-a608c7e2-51a6-4791-b00b-570fb8a353b0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.490s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.852907] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-866dfa01-544c-4b77-80ee-d1d6aa5af90d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.862910] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd8c9793-f7ce-403f-8360-597e8820df7e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.898013] env[61824]: DEBUG nova.compute.manager [req-0eb2c3ba-4dca-473c-a902-06a2306a6323 req-839c47e4-ae72-4293-8d1f-a4623849f4c2 service nova] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Detach interface failed, port_id=1a99ec03-ba39-4515-827d-6c5c7781f603, reason: Instance b1a770f9-71ea-4f80-a562-c08a313b5753 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 881.913100] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.957953] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f62f7f-fddb-4dc6-998c-11fb85813cc4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.967940] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a689a18-3c5e-47c6-a47a-042920e97791 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.009205] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5c5410-309d-41d4-bb33-55c67608e97a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.013344] env[61824]: INFO nova.compute.manager [-] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Took 1.51 seconds to deallocate network for instance. [ 882.022363] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.023186] env[61824]: DEBUG oslo_concurrency.lockutils [None req-02b1c763-45df-4193-9bba-ceb3c5fc18af tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.478s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.030783] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc69b37-f1e1-4f9e-8a83-3ead440010fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.039719] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275729, 'name': ReconfigVM_Task, 'duration_secs': 0.262595} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.048494] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 09432e72-c10a-4a7a-93af-af0d371fb871/09432e72-c10a-4a7a-93af-af0d371fb871.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.049626] env[61824]: DEBUG nova.compute.provider_tree [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.052139] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4eb6855-9cdf-4a3e-a9b2-3f4786992f3a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.057178] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 882.057178] env[61824]: value = "task-1275730" [ 882.057178] env[61824]: _type = "Task" [ 882.057178] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.066100] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275730, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.169967] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "2495fcb4-28cb-49c3-90fe-c84072466287" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.170168] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "2495fcb4-28cb-49c3-90fe-c84072466287" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.172080] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "2495fcb4-28cb-49c3-90fe-c84072466287-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.172080] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "2495fcb4-28cb-49c3-90fe-c84072466287-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.172080] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "2495fcb4-28cb-49c3-90fe-c84072466287-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.172977] env[61824]: INFO nova.compute.manager [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Terminating instance [ 882.174952] env[61824]: DEBUG nova.compute.manager [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.175161] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.176032] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db04cfbb-1ba8-478b-b0aa-026a1cdd306d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.184234] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.184510] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f50507c-55f0-4687-b667-151d3c96667d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.190890] env[61824]: DEBUG oslo_vmware.api [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 882.190890] env[61824]: value = "task-1275731" [ 882.190890] env[61824]: _type = "Task" [ 882.190890] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.200861] env[61824]: DEBUG oslo_vmware.api [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.240573] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.240843] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.241072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.241294] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.241465] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.243578] env[61824]: INFO nova.compute.manager [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Terminating instance [ 882.245285] env[61824]: DEBUG nova.compute.manager [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.245479] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.246395] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-360dba79-7fcf-40a1-8677-5313ab07f4ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.253784] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.254015] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e6232c8-e3a4-4863-b502-bfaf32bedac3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.260074] env[61824]: DEBUG oslo_vmware.api [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 882.260074] env[61824]: value = "task-1275732" [ 882.260074] env[61824]: _type = "Task" [ 882.260074] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.267632] env[61824]: DEBUG oslo_vmware.api [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275732, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.523984] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.567613] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275730, 'name': Rename_Task, 'duration_secs': 0.155461} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.567919] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 882.568239] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a37b46e-91e6-47d4-8b76-2b5d67e8f141 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.571677] env[61824]: ERROR nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [req-bc4b1218-0386-455f-8bdf-d070a0bb7d2f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8d9ece0e-729c-4d69-b741-d89969200860. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bc4b1218-0386-455f-8bdf-d070a0bb7d2f"}]} [ 882.575193] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 882.575193] env[61824]: value = "task-1275733" [ 882.575193] env[61824]: _type = "Task" [ 882.575193] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.583019] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.587691] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Refreshing inventories for resource provider 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 882.602121] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating ProviderTree inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 882.602415] env[61824]: DEBUG nova.compute.provider_tree [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.615139] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Refreshing aggregate associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, aggregates: None {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 882.635675] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Refreshing trait associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 882.701950] env[61824]: DEBUG oslo_vmware.api [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275731, 'name': PowerOffVM_Task, 'duration_secs': 0.386822} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.703986] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.703986] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.707696] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54baf3dd-ac75-415a-b00b-1a27c3be6551 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.771717] env[61824]: DEBUG oslo_vmware.api [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275732, 'name': PowerOffVM_Task, 'duration_secs': 0.345169} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.771998] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.772194] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.772463] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-689070f1-bbf0-4393-951f-210ff36d0c32 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.783188] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.783417] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.783603] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleting the datastore file [datastore2] 2495fcb4-28cb-49c3-90fe-c84072466287 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.784164] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11744e96-ac9a-42c9-a952-a363e01f7946 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.792288] env[61824]: DEBUG oslo_vmware.api [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 882.792288] env[61824]: value = "task-1275736" [ 882.792288] env[61824]: _type = "Task" [ 882.792288] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.799990] env[61824]: DEBUG oslo_vmware.api [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275736, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.840269] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.840592] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.840835] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleting the datastore file [datastore1] 01fd9faa-a0de-45b2-89bf-094c6d1ac155 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.841144] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10a78bad-979d-4172-86d7-ad4e1adef1e4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.847723] env[61824]: DEBUG oslo_vmware.api [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 882.847723] env[61824]: value = "task-1275737" [ 882.847723] env[61824]: _type = "Task" [ 882.847723] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.858278] env[61824]: DEBUG oslo_vmware.api [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275737, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.880977] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca8b58d-7cb7-4962-84d2-b5901db12438 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.888664] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e961b7fc-2153-40ef-95f8-6fb2250f3e15 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.920703] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0520d23f-95ad-4cab-af6e-c7432ac4db04 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.924601] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.924828] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.929217] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "beafe3e0-ff41-40f7-90cf-1d7977191180" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.929422] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.934310] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7808239-bf71-450d-a03e-ff04edba2b43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.951193] env[61824]: DEBUG nova.compute.provider_tree [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 883.086545] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275733, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.210695] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "interface-5569eb0a-fd42-4e2d-b805-bb40e3264a65-a608c7e2-51a6-4791-b00b-570fb8a353b0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.210695] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-5569eb0a-fd42-4e2d-b805-bb40e3264a65-a608c7e2-51a6-4791-b00b-570fb8a353b0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.305834] env[61824]: DEBUG oslo_vmware.api [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275736, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155792} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.306170] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.306449] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.306721] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.306980] env[61824]: INFO nova.compute.manager [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Took 1.13 seconds to destroy the instance on the hypervisor. [ 883.307350] env[61824]: DEBUG oslo.service.loopingcall [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.307622] env[61824]: DEBUG nova.compute.manager [-] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.307765] env[61824]: DEBUG nova.network.neutron [-] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 883.362485] env[61824]: DEBUG oslo_vmware.api [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275737, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145008} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.362927] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.363273] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.363569] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.363852] env[61824]: INFO nova.compute.manager [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Took 1.12 seconds to destroy the instance on the hypervisor. [ 883.364234] env[61824]: DEBUG oslo.service.loopingcall [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.364556] env[61824]: DEBUG nova.compute.manager [-] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.364720] env[61824]: DEBUG nova.network.neutron [-] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 883.374857] env[61824]: DEBUG nova.compute.manager [req-0ccd6fd5-3a08-484c-933d-e30cac04c02f req-2f9288a9-0700-44cb-be66-fa52b8cb692e service nova] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Received event network-vif-deleted-66cc8af8-5b27-44c8-b043-106fe1e38714 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.429747] env[61824]: INFO nova.compute.manager [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Detaching volume 779f9475-aa56-4fae-b35f-ce65c3cfc4be [ 883.434075] env[61824]: DEBUG nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 883.476119] env[61824]: INFO nova.virt.block_device [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Attempting to driver detach volume 779f9475-aa56-4fae-b35f-ce65c3cfc4be from mountpoint /dev/sdb [ 883.476119] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 883.476119] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274172', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'name': 'volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'serial': '779f9475-aa56-4fae-b35f-ce65c3cfc4be'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 883.476119] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6ab9bf-5358-46b7-a340-e37902c7f26c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.505014] env[61824]: DEBUG nova.scheduler.client.report [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updated inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 with generation 97 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 883.505367] env[61824]: DEBUG nova.compute.provider_tree [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating resource provider 8d9ece0e-729c-4d69-b741-d89969200860 generation from 97 to 98 during operation: update_inventory {{(pid=61824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 883.505614] env[61824]: DEBUG nova.compute.provider_tree [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 883.509617] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e1e40e-c729-4d0f-99ce-dc5e4c17b18b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.518297] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34812a67-db1d-4ef6-bdb4-9921577cf45c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.547318] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f70c12a-65dd-48c0-b419-efafa9f6ba7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.561705] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] The volume has not been displaced from its original location: [datastore1] volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be/volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 883.568812] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfiguring VM instance instance-00000042 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 883.569359] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f699dd84-d193-48b0-b4ee-46fdda0c02f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.595469] env[61824]: DEBUG oslo_vmware.api [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275733, 'name': PowerOnVM_Task, 'duration_secs': 0.546656} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.597637] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 883.597895] env[61824]: INFO nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Took 8.80 seconds to spawn the instance on the hypervisor. [ 883.598168] env[61824]: DEBUG nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 883.598542] env[61824]: DEBUG oslo_vmware.api [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 883.598542] env[61824]: value = "task-1275738" [ 883.598542] env[61824]: _type = "Task" [ 883.598542] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.599338] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bc9986-6f30-4cd6-9e01-f676f0d31985 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.614874] env[61824]: DEBUG oslo_vmware.api [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275738, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.706371] env[61824]: DEBUG nova.compute.manager [req-c0656f47-7d47-42e4-8f61-aaa0814b78f4 req-8dc1a18a-319f-479b-a11b-8dd76a99b020 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Received event network-vif-deleted-0660bee3-4b22-4474-ab43-1a3947ada0b1 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.706371] env[61824]: INFO nova.compute.manager [req-c0656f47-7d47-42e4-8f61-aaa0814b78f4 req-8dc1a18a-319f-479b-a11b-8dd76a99b020 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Neutron deleted interface 0660bee3-4b22-4474-ab43-1a3947ada0b1; detaching it from the instance and deleting it from the info cache [ 883.706661] env[61824]: DEBUG nova.network.neutron [req-c0656f47-7d47-42e4-8f61-aaa0814b78f4 req-8dc1a18a-319f-479b-a11b-8dd76a99b020 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.713313] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.713716] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.714811] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7522e89f-f092-4447-ba01-52682ce9e8c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.743204] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911c2cdc-1775-4502-bef3-5abb83270a99 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.777197] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Reconfiguring VM to detach interface {{(pid=61824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 883.777541] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47bcb268-a943-4035-b2c2-dd3aeb3b1008 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.796430] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 883.796430] env[61824]: value = "task-1275739" [ 883.796430] env[61824]: _type = "Task" [ 883.796430] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.804738] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.961472] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.014549] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.303s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.018051] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.741s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.020065] env[61824]: INFO nova.compute.claims [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.086373] env[61824]: DEBUG nova.network.neutron [-] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.087558] env[61824]: DEBUG nova.network.neutron [-] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.111654] env[61824]: DEBUG oslo_vmware.api [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275738, 'name': ReconfigVM_Task, 'duration_secs': 0.408422} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.111942] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfigured VM instance instance-00000042 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 884.120253] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c74d48d0-b2ef-47c3-8a6d-1deb4f49c4f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.131771] env[61824]: INFO nova.compute.manager [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Took 17.50 seconds to build instance. [ 884.136413] env[61824]: DEBUG oslo_vmware.api [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 884.136413] env[61824]: value = "task-1275740" [ 884.136413] env[61824]: _type = "Task" [ 884.136413] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.144489] env[61824]: DEBUG oslo_vmware.api [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275740, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.213367] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f6c1d48-e138-4e66-8264-cb3488296695 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.223395] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5367ca65-c5ee-4d18-85f0-beb4f8e6fd34 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.258038] env[61824]: DEBUG nova.compute.manager [req-c0656f47-7d47-42e4-8f61-aaa0814b78f4 req-8dc1a18a-319f-479b-a11b-8dd76a99b020 service nova] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Detach interface failed, port_id=0660bee3-4b22-4474-ab43-1a3947ada0b1, reason: Instance 2495fcb4-28cb-49c3-90fe-c84072466287 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 884.305328] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.530401] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dcc7fe39-a11c-4702-be16-b0813f2c6ef7 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 27.594s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.531227] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.679s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.531386] env[61824]: INFO nova.compute.manager [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Unshelving [ 884.590072] env[61824]: INFO nova.compute.manager [-] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Took 1.23 seconds to deallocate network for instance. [ 884.590425] env[61824]: INFO nova.compute.manager [-] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Took 1.28 seconds to deallocate network for instance. [ 884.633980] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e2c853d3-a245-4ecf-913e-2619507c691e tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.027s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.646392] env[61824]: DEBUG oslo_vmware.api [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275740, 'name': ReconfigVM_Task, 'duration_secs': 0.1287} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.646731] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274172', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'name': 'volume-779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': '779f9475-aa56-4fae-b35f-ce65c3cfc4be', 'serial': '779f9475-aa56-4fae-b35f-ce65c3cfc4be'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 884.807898] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.099340] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.101890] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.209822] env[61824]: DEBUG nova.objects.instance [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lazy-loading 'flavor' on Instance uuid cbded637-56b9-4048-b66e-1edf6311f435 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.236211] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759444f3-98eb-442c-9fed-54dd9e24649d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.245031] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b244a01-b57e-4e83-b205-73fe52af3259 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.277084] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39a72ef-e07d-48ea-96ec-a462eb8d956c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.284634] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622bc8ae-d0a3-4892-ad58-dfb62b6964fb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.297568] env[61824]: DEBUG nova.compute.provider_tree [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.306203] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.401310] env[61824]: DEBUG nova.compute.manager [req-7037a2b0-c78b-499a-a8c1-724052f02346 req-8be80dae-7fe6-4b55-b76f-7aba11afa4f2 service nova] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Received event network-vif-deleted-e0fb98c3-2363-41c7-b63f-f3550b0050be {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.541108] env[61824]: DEBUG nova.compute.utils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.804114] env[61824]: DEBUG nova.scheduler.client.report [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.813695] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.044943] env[61824]: INFO nova.virt.block_device [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Booting with volume 279b78fe-a845-4106-9b4f-158ad3979315 at /dev/sdb [ 886.079374] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2c310fbb-0347-444f-930c-d49115453360 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.088892] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79adf77c-a91c-4d98-8602-3ceb46d38f6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.117148] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-474fc956-5859-45d7-96de-c580433aaf37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.125033] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f099b9da-371c-4f07-b276-b0eb54b9e02a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.151571] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d113e8e-1a0d-4c07-87a4-b2a082b5da70 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.157567] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0932cef7-db2a-490d-81a8-fb4ba5aecc4b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.170207] env[61824]: DEBUG nova.virt.block_device [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating existing volume attachment record: 8763ed34-759e-4802-8cb1-cace5a9e956f {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 886.181604] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.219377] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83959e6-4448-46a5-b342-5d54c1dd99bf tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.294s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.221761] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.039s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.284635] env[61824]: DEBUG nova.compute.manager [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Stashing vm_state: active {{(pid=61824) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 886.309023] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.309672] env[61824]: DEBUG nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 886.312050] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.312323] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.022s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.312835] env[61824]: DEBUG nova.objects.instance [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 886.723171] env[61824]: INFO nova.compute.manager [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Detaching volume fd8092a8-d5fb-41a2-b612-00a7cff5fbe7 [ 886.758940] env[61824]: INFO nova.virt.block_device [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Attempting to driver detach volume fd8092a8-d5fb-41a2-b612-00a7cff5fbe7 from mountpoint /dev/sdc [ 886.759219] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 886.759411] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274178', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'name': 'volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'serial': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 886.760298] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80b3192-6c2d-4cf7-b62b-04899f72c61d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.782885] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311436a2-2ef7-4a46-890d-f2021b208e54 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.793031] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c01eac17-c552-4dfd-b596-b7cd12290eef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.813878] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.817261] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654b3e1c-7332-482b-ad08-2c2a8a83a6c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.820625] env[61824]: DEBUG nova.compute.utils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 886.824528] env[61824]: DEBUG nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 886.824707] env[61824]: DEBUG nova.network.neutron [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 886.842170] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.842372] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] The volume has not been displaced from its original location: [datastore2] volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7/volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 886.847630] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfiguring VM instance instance-00000042 to detach disk 2002 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 886.848954] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24b7d2e5-b10a-407e-91af-75e7645aeeee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.863664] env[61824]: DEBUG nova.policy [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a6526db8b7343b3b9c182d266f5db64', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9edafb45f7124fb89adfced8c0a07f2b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 886.866966] env[61824]: DEBUG oslo_vmware.api [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 886.866966] env[61824]: value = "task-1275744" [ 886.866966] env[61824]: _type = "Task" [ 886.866966] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.877140] env[61824]: DEBUG oslo_vmware.api [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275744, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.237891] env[61824]: DEBUG nova.network.neutron [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Successfully created port: adefa466-eb42-422d-ade5-3f7cf6e305ef {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.323801] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.325456] env[61824]: DEBUG nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 887.328799] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8bc99c97-224f-446b-8c0c-61abe1711ddd tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.329876] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.802s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.331300] env[61824]: INFO nova.compute.claims [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.376464] env[61824]: DEBUG oslo_vmware.api [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275744, 'name': ReconfigVM_Task, 'duration_secs': 0.252422} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.377372] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Reconfigured VM instance instance-00000042 to detach disk 2002 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 887.382725] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cbcb6bf-e235-49a1-b5a2-82e84db94fc5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.399874] env[61824]: DEBUG oslo_vmware.api [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 887.399874] env[61824]: value = "task-1275745" [ 887.399874] env[61824]: _type = "Task" [ 887.399874] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.407979] env[61824]: DEBUG oslo_vmware.api [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275745, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.824212] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.910431] env[61824]: DEBUG oslo_vmware.api [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275745, 'name': ReconfigVM_Task, 'duration_secs': 0.123443} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.910431] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274178', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'name': 'volume-fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'cbded637-56b9-4048-b66e-1edf6311f435', 'attached_at': '', 'detached_at': '', 'volume_id': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7', 'serial': 'fd8092a8-d5fb-41a2-b612-00a7cff5fbe7'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 888.325618] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.335877] env[61824]: DEBUG nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 888.361310] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.361575] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.361737] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.361926] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.362096] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.362255] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.362464] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.362627] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.362793] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.362955] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.363152] env[61824]: DEBUG nova.virt.hardware [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.364181] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a624a8-7581-4f89-8a5c-0413408f1efd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.372326] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f616d2-beff-42a3-b915-0d349b5fb5fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.451952] env[61824]: DEBUG nova.objects.instance [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lazy-loading 'flavor' on Instance uuid cbded637-56b9-4048-b66e-1edf6311f435 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.571480] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5d2678-e0d6-4eab-bd24-6a641d2721a2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.578879] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed475d2d-7ff8-456d-932a-17ddd713e445 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.609727] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd204d0c-580f-46a9-b59d-02a661de66e7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.617265] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc5b1518-4299-4aca-a458-4296b624edc9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.631726] env[61824]: DEBUG nova.compute.provider_tree [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.825787] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.853779] env[61824]: DEBUG nova.compute.manager [req-3429f489-d173-456a-b119-9c6eaa14c32c req-fe3d7484-5f61-4c05-89ec-e81812a4afff service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Received event network-vif-plugged-adefa466-eb42-422d-ade5-3f7cf6e305ef {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.854011] env[61824]: DEBUG oslo_concurrency.lockutils [req-3429f489-d173-456a-b119-9c6eaa14c32c req-fe3d7484-5f61-4c05-89ec-e81812a4afff service nova] Acquiring lock "9f031313-766b-4696-9bc6-75ced078bfde-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.854243] env[61824]: DEBUG oslo_concurrency.lockutils [req-3429f489-d173-456a-b119-9c6eaa14c32c req-fe3d7484-5f61-4c05-89ec-e81812a4afff service nova] Lock "9f031313-766b-4696-9bc6-75ced078bfde-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.854422] env[61824]: DEBUG oslo_concurrency.lockutils [req-3429f489-d173-456a-b119-9c6eaa14c32c req-fe3d7484-5f61-4c05-89ec-e81812a4afff service nova] Lock "9f031313-766b-4696-9bc6-75ced078bfde-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.854587] env[61824]: DEBUG nova.compute.manager [req-3429f489-d173-456a-b119-9c6eaa14c32c req-fe3d7484-5f61-4c05-89ec-e81812a4afff service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] No waiting events found dispatching network-vif-plugged-adefa466-eb42-422d-ade5-3f7cf6e305ef {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.854757] env[61824]: WARNING nova.compute.manager [req-3429f489-d173-456a-b119-9c6eaa14c32c req-fe3d7484-5f61-4c05-89ec-e81812a4afff service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Received unexpected event network-vif-plugged-adefa466-eb42-422d-ade5-3f7cf6e305ef for instance with vm_state building and task_state spawning. [ 888.945919] env[61824]: DEBUG nova.network.neutron [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Successfully updated port: adefa466-eb42-422d-ade5-3f7cf6e305ef {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.134952] env[61824]: DEBUG nova.scheduler.client.report [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.326801] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.452275] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "refresh_cache-9f031313-766b-4696-9bc6-75ced078bfde" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.452522] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquired lock "refresh_cache-9f031313-766b-4696-9bc6-75ced078bfde" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.452699] env[61824]: DEBUG nova.network.neutron [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.461130] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3af3901e-f45e-4435-90ba-fc1b87305ab6 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.240s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.640696] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.640696] env[61824]: DEBUG nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.643348] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.730s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.645619] env[61824]: INFO nova.compute.claims [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.827798] env[61824]: DEBUG oslo_vmware.api [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275739, 'name': ReconfigVM_Task, 'duration_secs': 5.741619} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.828163] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.828292] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Reconfigured VM to detach interface {{(pid=61824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 889.986218] env[61824]: DEBUG nova.network.neutron [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 890.121381] env[61824]: DEBUG nova.network.neutron [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Updating instance_info_cache with network_info: [{"id": "adefa466-eb42-422d-ade5-3f7cf6e305ef", "address": "fa:16:3e:e3:f9:68", "network": {"id": "d3046df1-15f3-46df-a78b-d0675c835194", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-934659660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9edafb45f7124fb89adfced8c0a07f2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a766368-f5a3-472c-af56-9cfca63012ae", "external-id": "nsx-vlan-transportzone-518", "segmentation_id": 518, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadefa466-eb", "ovs_interfaceid": "adefa466-eb42-422d-ade5-3f7cf6e305ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.149368] env[61824]: DEBUG nova.compute.utils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.152801] env[61824]: DEBUG nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 890.152979] env[61824]: DEBUG nova.network.neutron [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 890.191528] env[61824]: DEBUG nova.policy [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1eb545ac2cf4f0d9ccddb040b17372b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e756c6817698464a9da391d4dd364946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 890.440182] env[61824]: DEBUG nova.network.neutron [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Successfully created port: 734e041a-690b-4a80-9b58-6c6bcf162752 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.567774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.567774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.567774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "cbded637-56b9-4048-b66e-1edf6311f435-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.567774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.567774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.571074] env[61824]: INFO nova.compute.manager [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Terminating instance [ 890.576026] env[61824]: DEBUG nova.compute.manager [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 890.576026] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 890.576026] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a60bddb-4679-4f20-8516-ae2840669591 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.585298] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.585298] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06808eff-64fa-4c82-99d2-49d8b348b648 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.596272] env[61824]: DEBUG oslo_vmware.api [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 890.596272] env[61824]: value = "task-1275747" [ 890.596272] env[61824]: _type = "Task" [ 890.596272] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.602303] env[61824]: DEBUG oslo_vmware.api [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275747, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.624250] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Releasing lock "refresh_cache-9f031313-766b-4696-9bc6-75ced078bfde" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.624632] env[61824]: DEBUG nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Instance network_info: |[{"id": "adefa466-eb42-422d-ade5-3f7cf6e305ef", "address": "fa:16:3e:e3:f9:68", "network": {"id": "d3046df1-15f3-46df-a78b-d0675c835194", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-934659660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9edafb45f7124fb89adfced8c0a07f2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a766368-f5a3-472c-af56-9cfca63012ae", "external-id": "nsx-vlan-transportzone-518", "segmentation_id": 518, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadefa466-eb", "ovs_interfaceid": "adefa466-eb42-422d-ade5-3f7cf6e305ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 890.625103] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:f9:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a766368-f5a3-472c-af56-9cfca63012ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'adefa466-eb42-422d-ade5-3f7cf6e305ef', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.634095] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Creating folder: Project (9edafb45f7124fb89adfced8c0a07f2b). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.634432] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7dd01f8-c996-4533-bed5-e03d3457da01 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.644891] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Created folder: Project (9edafb45f7124fb89adfced8c0a07f2b) in parent group-v274074. [ 890.645196] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Creating folder: Instances. Parent ref: group-v274182. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.645499] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e0aca53-2ef0-4384-9e39-2a1ccd72dd87 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.654753] env[61824]: DEBUG nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.657513] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Created folder: Instances in parent group-v274182. [ 890.657821] env[61824]: DEBUG oslo.service.loopingcall [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.660995] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.661489] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-72868861-8f96-45f0-93a6-b85858424b4e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.684945] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.684945] env[61824]: value = "task-1275750" [ 890.684945] env[61824]: _type = "Task" [ 890.684945] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.697584] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275750, 'name': CreateVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.883202] env[61824]: DEBUG nova.compute.manager [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Received event network-changed-adefa466-eb42-422d-ade5-3f7cf6e305ef {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.883569] env[61824]: DEBUG nova.compute.manager [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Refreshing instance network info cache due to event network-changed-adefa466-eb42-422d-ade5-3f7cf6e305ef. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 890.883943] env[61824]: DEBUG oslo_concurrency.lockutils [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] Acquiring lock "refresh_cache-9f031313-766b-4696-9bc6-75ced078bfde" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.884212] env[61824]: DEBUG oslo_concurrency.lockutils [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] Acquired lock "refresh_cache-9f031313-766b-4696-9bc6-75ced078bfde" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.884511] env[61824]: DEBUG nova.network.neutron [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Refreshing network info cache for port adefa466-eb42-422d-ade5-3f7cf6e305ef {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 890.950704] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b4aeb5-fc0d-4ccd-99b4-64516f61ab88 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.958777] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e665810-051c-4bb6-bc90-253c7c74360f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.990436] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-badd0d56-8f92-4252-99e6-7b537d9f416f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.997742] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-062d14c9-00e8-42ab-94d7-1f402ca1c73e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.011759] env[61824]: DEBUG nova.compute.provider_tree [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.102232] env[61824]: DEBUG oslo_vmware.api [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275747, 'name': PowerOffVM_Task, 'duration_secs': 0.311028} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.102522] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.102694] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.102940] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0061c38a-8bf0-49b3-b3b8-5aa09e9fd656 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.165053] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.165053] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.165273] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Deleting the datastore file [datastore2] cbded637-56b9-4048-b66e-1edf6311f435 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.165489] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3abd2ca8-14c7-411b-b534-b991c1b10981 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.171204] env[61824]: DEBUG oslo_vmware.api [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for the task: (returnval){ [ 891.171204] env[61824]: value = "task-1275752" [ 891.171204] env[61824]: _type = "Task" [ 891.171204] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.180152] env[61824]: DEBUG oslo_vmware.api [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275752, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.193615] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275750, 'name': CreateVM_Task, 'duration_secs': 0.427292} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.194452] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.195163] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.195368] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.195746] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.196281] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a7af76c-8763-4aa1-9a6d-b6f2801e9e5f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.200827] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 891.200827] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]527c3d14-83cc-e352-18a7-75ff540a7218" [ 891.200827] env[61824]: _type = "Task" [ 891.200827] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.205028] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.205028] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.205028] env[61824]: DEBUG nova.network.neutron [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.210590] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527c3d14-83cc-e352-18a7-75ff540a7218, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.516442] env[61824]: DEBUG nova.scheduler.client.report [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 891.666595] env[61824]: DEBUG nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.688476] env[61824]: DEBUG oslo_vmware.api [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Task: {'id': task-1275752, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141143} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.690769] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 891.690869] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 891.691017] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 891.691288] env[61824]: INFO nova.compute.manager [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Took 1.12 seconds to destroy the instance on the hypervisor. [ 891.691431] env[61824]: DEBUG oslo.service.loopingcall [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 891.691824] env[61824]: DEBUG nova.compute.manager [-] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 891.691923] env[61824]: DEBUG nova.network.neutron [-] [instance: cbded637-56b9-4048-b66e-1edf6311f435] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 891.698853] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.699088] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.699255] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.699486] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.699623] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.699737] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.699942] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.700141] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.700323] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.700494] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.700681] env[61824]: DEBUG nova.virt.hardware [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.701501] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a87ef02-a22f-4cc5-8fab-485cd1bcfe85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.713083] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53df0f40-931b-4d6e-9030-e135f57babcb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.720206] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527c3d14-83cc-e352-18a7-75ff540a7218, 'name': SearchDatastore_Task, 'duration_secs': 0.017284} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.720789] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.721034] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.721278] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.721432] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.721612] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.721859] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cafbabc1-eb23-4e37-884f-5df4ff07cc0e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.739590] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.739862] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.740901] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aece32d7-a7bf-41d7-9a8f-cad5a223cddb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.747745] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 891.747745] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52b4c2bb-db1f-81d8-d281-59a4b32d35cb" [ 891.747745] env[61824]: _type = "Task" [ 891.747745] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.764292] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b4c2bb-db1f-81d8-d281-59a4b32d35cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.778351] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.955648] env[61824]: DEBUG nova.network.neutron [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Successfully updated port: 734e041a-690b-4a80-9b58-6c6bcf162752 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.024506] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.025063] env[61824]: DEBUG nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.028723] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.006s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.028957] env[61824]: DEBUG nova.objects.instance [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lazy-loading 'resources' on Instance uuid b1a770f9-71ea-4f80-a562-c08a313b5753 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.042563] env[61824]: DEBUG nova.compute.manager [req-fef0ddee-90ba-46e6-9416-399cfe684155 req-6d2e23e1-4680-4524-865d-54b5b0839cf0 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Received event network-vif-plugged-734e041a-690b-4a80-9b58-6c6bcf162752 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.042798] env[61824]: DEBUG oslo_concurrency.lockutils [req-fef0ddee-90ba-46e6-9416-399cfe684155 req-6d2e23e1-4680-4524-865d-54b5b0839cf0 service nova] Acquiring lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.043041] env[61824]: DEBUG oslo_concurrency.lockutils [req-fef0ddee-90ba-46e6-9416-399cfe684155 req-6d2e23e1-4680-4524-865d-54b5b0839cf0 service nova] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.043222] env[61824]: DEBUG oslo_concurrency.lockutils [req-fef0ddee-90ba-46e6-9416-399cfe684155 req-6d2e23e1-4680-4524-865d-54b5b0839cf0 service nova] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.043462] env[61824]: DEBUG nova.compute.manager [req-fef0ddee-90ba-46e6-9416-399cfe684155 req-6d2e23e1-4680-4524-865d-54b5b0839cf0 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] No waiting events found dispatching network-vif-plugged-734e041a-690b-4a80-9b58-6c6bcf162752 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.043566] env[61824]: WARNING nova.compute.manager [req-fef0ddee-90ba-46e6-9416-399cfe684155 req-6d2e23e1-4680-4524-865d-54b5b0839cf0 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Received unexpected event network-vif-plugged-734e041a-690b-4a80-9b58-6c6bcf162752 for instance with vm_state building and task_state spawning. [ 892.139221] env[61824]: DEBUG nova.network.neutron [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Updated VIF entry in instance network info cache for port adefa466-eb42-422d-ade5-3f7cf6e305ef. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 892.139602] env[61824]: DEBUG nova.network.neutron [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Updating instance_info_cache with network_info: [{"id": "adefa466-eb42-422d-ade5-3f7cf6e305ef", "address": "fa:16:3e:e3:f9:68", "network": {"id": "d3046df1-15f3-46df-a78b-d0675c835194", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-934659660-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9edafb45f7124fb89adfced8c0a07f2b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a766368-f5a3-472c-af56-9cfca63012ae", "external-id": "nsx-vlan-transportzone-518", "segmentation_id": 518, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadefa466-eb", "ovs_interfaceid": "adefa466-eb42-422d-ade5-3f7cf6e305ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.258641] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b4c2bb-db1f-81d8-d281-59a4b32d35cb, 'name': SearchDatastore_Task, 'duration_secs': 0.012696} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.259486] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30e47230-0b04-4e28-aeef-212b122e2726 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.264921] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 892.264921] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52556a18-d620-7ddf-4511-8b8a1c5f169f" [ 892.264921] env[61824]: _type = "Task" [ 892.264921] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.272518] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52556a18-d620-7ddf-4511-8b8a1c5f169f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.457499] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "refresh_cache-07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.457693] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "refresh_cache-07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.457768] env[61824]: DEBUG nova.network.neutron [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 892.531345] env[61824]: DEBUG nova.compute.utils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 892.532764] env[61824]: DEBUG nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 892.532936] env[61824]: DEBUG nova.network.neutron [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 892.642699] env[61824]: DEBUG oslo_concurrency.lockutils [req-f785b02a-fc8c-4e4d-8108-72c882d95a91 req-ff0ad223-1e86-447f-aa8e-0763ac62e3b7 service nova] Releasing lock "refresh_cache-9f031313-766b-4696-9bc6-75ced078bfde" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.720932] env[61824]: INFO nova.network.neutron [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Port a608c7e2-51a6-4791-b00b-570fb8a353b0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 892.721460] env[61824]: DEBUG nova.network.neutron [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.723517] env[61824]: DEBUG nova.policy [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a975aca3353469a948c52b035e17fcc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '502b41859ba843838d2188c82f70ae08', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 892.777885] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52556a18-d620-7ddf-4511-8b8a1c5f169f, 'name': SearchDatastore_Task, 'duration_secs': 0.040888} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.780273] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.780549] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 9f031313-766b-4696-9bc6-75ced078bfde/9f031313-766b-4696-9bc6-75ced078bfde.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.781603] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c39ab4b5-876f-458f-8a70-db5892e93427 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.787260] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 892.787260] env[61824]: value = "task-1275753" [ 892.787260] env[61824]: _type = "Task" [ 892.787260] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.796261] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275753, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.802908] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-974fd20a-69fc-4875-a454-e3b28b4dd140 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.809792] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f9d656-a98a-4005-94ef-9e06a2270dd4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.841365] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08996b1a-5fb7-499d-9dcd-960501e7c3c9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.850199] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a552e1a8-719d-46d1-8f3c-24aade823b8c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.865246] env[61824]: DEBUG nova.compute.provider_tree [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 893.036062] env[61824]: DEBUG nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 893.074179] env[61824]: DEBUG nova.network.neutron [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 893.229978] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.297441] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275753, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439839} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.297721] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 9f031313-766b-4696-9bc6-75ced078bfde/9f031313-766b-4696-9bc6-75ced078bfde.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.297936] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.298208] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b9c1a48-4843-4618-a60f-eac4b7bc6595 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.304632] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 893.304632] env[61824]: value = "task-1275754" [ 893.304632] env[61824]: _type = "Task" [ 893.304632] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.312453] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275754, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.368993] env[61824]: DEBUG nova.scheduler.client.report [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.717539] env[61824]: DEBUG nova.network.neutron [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Updating instance_info_cache with network_info: [{"id": "734e041a-690b-4a80-9b58-6c6bcf162752", "address": "fa:16:3e:03:3e:76", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap734e041a-69", "ovs_interfaceid": "734e041a-690b-4a80-9b58-6c6bcf162752", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.734921] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3447f58a-2946-4937-95ad-ff0e398db2b4 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-5569eb0a-fd42-4e2d-b805-bb40e3264a65-a608c7e2-51a6-4791-b00b-570fb8a353b0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.524s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.815083] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054151} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.815364] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.817145] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88007a6a-7590-4798-ac4d-3391ee8317b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.821821] env[61824]: DEBUG nova.compute.manager [req-9fc7e776-6040-49ce-85df-c7f2c1e5063c req-b16ffc7c-f21a-4bf6-8194-fd187ea0af59 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Received event network-vif-deleted-ee9b9520-3d4f-48fc-bb6d-04a43506c1aa {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 893.822000] env[61824]: INFO nova.compute.manager [req-9fc7e776-6040-49ce-85df-c7f2c1e5063c req-b16ffc7c-f21a-4bf6-8194-fd187ea0af59 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Neutron deleted interface ee9b9520-3d4f-48fc-bb6d-04a43506c1aa; detaching it from the instance and deleting it from the info cache [ 893.822193] env[61824]: DEBUG nova.network.neutron [req-9fc7e776-6040-49ce-85df-c7f2c1e5063c req-b16ffc7c-f21a-4bf6-8194-fd187ea0af59 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.856259] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 9f031313-766b-4696-9bc6-75ced078bfde/9f031313-766b-4696-9bc6-75ced078bfde.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.857087] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf9f7c6a-595f-442b-971e-a9af4b4b9e49 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.876055] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.878480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.355s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.878671] env[61824]: DEBUG nova.objects.instance [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lazy-loading 'resources' on Instance uuid c6552176-1128-473b-ac4a-f3b55428fd9e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.885861] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 893.885861] env[61824]: value = "task-1275755" [ 893.885861] env[61824]: _type = "Task" [ 893.885861] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.896803] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275755, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.900746] env[61824]: INFO nova.scheduler.client.report [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Deleted allocations for instance b1a770f9-71ea-4f80-a562-c08a313b5753 [ 893.945752] env[61824]: DEBUG nova.network.neutron [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Successfully created port: 41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.053573] env[61824]: DEBUG nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 894.079671] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.079868] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.079989] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.080284] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.080444] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.080634] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.080854] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.081036] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.081209] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.081381] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.081560] env[61824]: DEBUG nova.virt.hardware [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.082733] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc61fe8-6643-49ff-8688-5c7fa1ac9ac1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.090821] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329c098d-6161-4479-97ee-0f8819b9bf91 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.222087] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "refresh_cache-07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.222087] env[61824]: DEBUG nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Instance network_info: |[{"id": "734e041a-690b-4a80-9b58-6c6bcf162752", "address": "fa:16:3e:03:3e:76", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap734e041a-69", "ovs_interfaceid": "734e041a-690b-4a80-9b58-6c6bcf162752", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 894.222431] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:3e:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '734e041a-690b-4a80-9b58-6c6bcf162752', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.232523] env[61824]: DEBUG oslo.service.loopingcall [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.232523] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.232523] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8de1920-aee5-40fb-a27d-37d75b3964b6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.251489] env[61824]: DEBUG nova.network.neutron [-] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.253120] env[61824]: DEBUG nova.compute.manager [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Received event network-changed-734e041a-690b-4a80-9b58-6c6bcf162752 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.253312] env[61824]: DEBUG nova.compute.manager [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Refreshing instance network info cache due to event network-changed-734e041a-690b-4a80-9b58-6c6bcf162752. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 894.253527] env[61824]: DEBUG oslo_concurrency.lockutils [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] Acquiring lock "refresh_cache-07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.253673] env[61824]: DEBUG oslo_concurrency.lockutils [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] Acquired lock "refresh_cache-07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.253831] env[61824]: DEBUG nova.network.neutron [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Refreshing network info cache for port 734e041a-690b-4a80-9b58-6c6bcf162752 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 894.260807] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.260807] env[61824]: value = "task-1275756" [ 894.260807] env[61824]: _type = "Task" [ 894.260807] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.273061] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275756, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.344857] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba2a3766-d101-4610-bf48-6b1c07d6cde8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.354778] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f905b4-081a-4654-afe4-e0c156b9440f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.390474] env[61824]: DEBUG nova.compute.manager [req-9fc7e776-6040-49ce-85df-c7f2c1e5063c req-b16ffc7c-f21a-4bf6-8194-fd187ea0af59 service nova] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Detach interface failed, port_id=ee9b9520-3d4f-48fc-bb6d-04a43506c1aa, reason: Instance cbded637-56b9-4048-b66e-1edf6311f435 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 894.400255] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275755, 'name': ReconfigVM_Task, 'duration_secs': 0.27146} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.400928] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 9f031313-766b-4696-9bc6-75ced078bfde/9f031313-766b-4696-9bc6-75ced078bfde.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.401573] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96985f3c-18a2-44a6-9112-058a9ba09f85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.408832] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 894.408832] env[61824]: value = "task-1275757" [ 894.408832] env[61824]: _type = "Task" [ 894.408832] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.409427] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8e0edb70-4b60-4497-b758-f4dedb4f0e38 tempest-ImagesNegativeTestJSON-1857205228 tempest-ImagesNegativeTestJSON-1857205228-project-member] Lock "b1a770f9-71ea-4f80-a562-c08a313b5753" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.194s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.423019] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275757, 'name': Rename_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.620877] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc256dc-47a3-42b1-97c8-bc061c72da2c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.628625] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236e6a93-1774-4b8d-83c0-227e696ece7c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.660598] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "interface-aa9c9eb2-b411-4eca-9f8d-3df377ffe351-a608c7e2-51a6-4791-b00b-570fb8a353b0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.660947] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-aa9c9eb2-b411-4eca-9f8d-3df377ffe351-a608c7e2-51a6-4791-b00b-570fb8a353b0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.661364] env[61824]: DEBUG nova.objects.instance [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lazy-loading 'flavor' on Instance uuid aa9c9eb2-b411-4eca-9f8d-3df377ffe351 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 894.663705] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a235444e-8066-4b1c-a1b6-4e15eed7204d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.673751] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbe54b8-3dbf-4303-86b8-dc7e3f8c8057 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.690059] env[61824]: DEBUG nova.compute.provider_tree [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.755969] env[61824]: INFO nova.compute.manager [-] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Took 3.06 seconds to deallocate network for instance. [ 894.773327] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275756, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.922764] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275757, 'name': Rename_Task, 'duration_secs': 0.329959} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.923382] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.923643] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7aaa7743-c1c4-4d58-a0c3-6b2c28326685 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.930245] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 894.930245] env[61824]: value = "task-1275758" [ 894.930245] env[61824]: _type = "Task" [ 894.930245] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.939625] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.194048] env[61824]: DEBUG nova.scheduler.client.report [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.267748] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.270152] env[61824]: DEBUG nova.network.neutron [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Updated VIF entry in instance network info cache for port 734e041a-690b-4a80-9b58-6c6bcf162752. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 895.270602] env[61824]: DEBUG nova.network.neutron [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Updating instance_info_cache with network_info: [{"id": "734e041a-690b-4a80-9b58-6c6bcf162752", "address": "fa:16:3e:03:3e:76", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap734e041a-69", "ovs_interfaceid": "734e041a-690b-4a80-9b58-6c6bcf162752", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.275109] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275756, 'name': CreateVM_Task, 'duration_secs': 0.589834} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.275462] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.276111] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.276320] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.276644] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.276912] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd6ea434-f203-49c1-b1e3-aa4a04219281 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.281587] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 895.281587] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5263fcb8-5322-a8d9-92dc-6bc5b46aa465" [ 895.281587] env[61824]: _type = "Task" [ 895.281587] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.291441] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5263fcb8-5322-a8d9-92dc-6bc5b46aa465, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.365839] env[61824]: DEBUG nova.objects.instance [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lazy-loading 'pci_requests' on Instance uuid aa9c9eb2-b411-4eca-9f8d-3df377ffe351 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.441730] env[61824]: DEBUG oslo_vmware.api [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275758, 'name': PowerOnVM_Task, 'duration_secs': 0.452729} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.442014] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.442221] env[61824]: INFO nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Took 7.11 seconds to spawn the instance on the hypervisor. [ 895.442402] env[61824]: DEBUG nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.443173] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca067ad-b36f-4a51-b897-2d803e5ebb58 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.679645] env[61824]: DEBUG nova.network.neutron [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Successfully updated port: 41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.698932] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.701392] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.740s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.703204] env[61824]: INFO nova.compute.claims [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.752144] env[61824]: INFO nova.scheduler.client.report [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleted allocations for instance c6552176-1128-473b-ac4a-f3b55428fd9e [ 895.777607] env[61824]: DEBUG oslo_concurrency.lockutils [req-36a95772-371e-4ad8-8f63-6acbb41684d8 req-111a1c9f-adf7-4273-a072-f63b1785b3b4 service nova] Releasing lock "refresh_cache-07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.793221] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5263fcb8-5322-a8d9-92dc-6bc5b46aa465, 'name': SearchDatastore_Task, 'duration_secs': 0.03091} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.793541] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.793790] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.794136] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.794207] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.794378] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.794663] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8bc2a8a-bb89-46ae-93d2-3bff3815b4ba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.803752] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.803940] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.804705] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-880a4e92-6656-45af-a5b3-69f1c7f4fe4e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.810760] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 895.810760] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5218e906-31ec-e54d-5db6-7e7e2c1823d3" [ 895.810760] env[61824]: _type = "Task" [ 895.810760] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.818752] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5218e906-31ec-e54d-5db6-7e7e2c1823d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.868987] env[61824]: DEBUG nova.objects.base [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 895.869249] env[61824]: DEBUG nova.network.neutron [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 895.960486] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.960654] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing instance network info cache due to event network-changed-9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 895.960824] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Acquiring lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.960919] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Acquired lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.961105] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Refreshing network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 895.963892] env[61824]: INFO nova.compute.manager [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Took 16.74 seconds to build instance. [ 895.967164] env[61824]: DEBUG nova.policy [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '316c089785a44aa897040f3c0ef7a071', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0924bfa41b5e42d68818557d2b0ace7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 896.182911] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.183099] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.183266] env[61824]: DEBUG nova.network.neutron [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.260558] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a1a91963-6087-44eb-bf54-3c6b88ee9d50 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "c6552176-1128-473b-ac4a-f3b55428fd9e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.051s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.321471] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5218e906-31ec-e54d-5db6-7e7e2c1823d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008002} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.322210] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f4dd316-2368-4878-8998-c8b1fa07752b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.327748] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 896.327748] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d86a4d-d8e6-abc2-7b72-bddad7fe1664" [ 896.327748] env[61824]: _type = "Task" [ 896.327748] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.336082] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d86a4d-d8e6-abc2-7b72-bddad7fe1664, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.467740] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d90d2ffa-c0c4-4ec1-9c58-03ed7ffd6ec1 tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "9f031313-766b-4696-9bc6-75ced078bfde" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.564s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.728155] env[61824]: DEBUG nova.network.neutron [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 896.732735] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updated VIF entry in instance network info cache for port 9baa0723-69a3-42a4-86c5-38ca3d336970. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 896.733081] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [{"id": "9baa0723-69a3-42a4-86c5-38ca3d336970", "address": "fa:16:3e:3b:98:59", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9baa0723-69", "ovs_interfaceid": "9baa0723-69a3-42a4-86c5-38ca3d336970", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.840617] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d86a4d-d8e6-abc2-7b72-bddad7fe1664, 'name': SearchDatastore_Task, 'duration_secs': 0.009557} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.843178] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.843457] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5/07973a77-0fe8-4bc9-b8c5-e693d7a63ef5.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.843931] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a69bb06c-83c1-4ca8-b2c6-5bb04046c4e3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.850318] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 896.850318] env[61824]: value = "task-1275759" [ 896.850318] env[61824]: _type = "Task" [ 896.850318] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.859680] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275759, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.934625] env[61824]: DEBUG nova.network.neutron [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Updating instance_info_cache with network_info: [{"id": "41458d47-a6e3-44c2-a378-9968713e2739", "address": "fa:16:3e:96:67:50", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41458d47-a6", "ovs_interfaceid": "41458d47-a6e3-44c2-a378-9968713e2739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.987968] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886717ca-8feb-42e1-94e1-2be217154a5b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.998899] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a11e2ce-393e-4293-aab5-57d11c609279 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.034681] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9c25fa-78dd-4402-86d2-d4a17f104234 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.043442] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4349b820-7878-4e2d-9a4d-b9dd3a871e92 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.061825] env[61824]: DEBUG nova.compute.provider_tree [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.235478] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Releasing lock "refresh_cache-5569eb0a-fd42-4e2d-b805-bb40e3264a65" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.235772] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.235948] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing instance network info cache due to event network-changed-bd35af69-df49-4c0a-a798-57bd9c8f6fe4. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.236230] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.236382] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.236724] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing network info cache for port bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 897.361362] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275759, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439884} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.361704] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5/07973a77-0fe8-4bc9-b8c5-e693d7a63ef5.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.361949] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.362242] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79253df0-2af4-432d-8ce0-9be41f7f1617 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.369023] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 897.369023] env[61824]: value = "task-1275760" [ 897.369023] env[61824]: _type = "Task" [ 897.369023] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.376472] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275760, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.437626] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.437968] env[61824]: DEBUG nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Instance network_info: |[{"id": "41458d47-a6e3-44c2-a378-9968713e2739", "address": "fa:16:3e:96:67:50", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41458d47-a6", "ovs_interfaceid": "41458d47-a6e3-44c2-a378-9968713e2739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.438486] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:67:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa410d21-2141-45bb-8d0b-16c77304605f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '41458d47-a6e3-44c2-a378-9968713e2739', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.447464] env[61824]: DEBUG oslo.service.loopingcall [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.447464] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.447464] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49953307-5655-441b-9a3a-0ff4681ae249 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.467482] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.467482] env[61824]: value = "task-1275761" [ 897.467482] env[61824]: _type = "Task" [ 897.467482] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.475496] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275761, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.566364] env[61824]: DEBUG nova.scheduler.client.report [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.682239] env[61824]: DEBUG nova.network.neutron [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Successfully updated port: a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 897.880817] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275760, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066619} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.881144] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 897.881998] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec71d31d-6bd0-4781-b0e8-92e64156037a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.909393] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5/07973a77-0fe8-4bc9-b8c5-e693d7a63ef5.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.909854] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25a077bb-7f68-47b3-8d06-a0db377ab49b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.933248] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 897.933248] env[61824]: value = "task-1275762" [ 897.933248] env[61824]: _type = "Task" [ 897.933248] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.943203] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275762, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.977396] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275761, 'name': CreateVM_Task, 'duration_secs': 0.392439} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.979964] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 897.980682] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.980861] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.981212] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 897.981781] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33b63cca-9b84-45ee-81d6-17721a71d931 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.986514] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 897.986514] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52244572-a2c7-79e1-55ee-edc00ca30a77" [ 897.986514] env[61824]: _type = "Task" [ 897.986514] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.989258] env[61824]: DEBUG nova.compute.manager [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-vif-plugged-a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.989778] env[61824]: DEBUG oslo_concurrency.lockutils [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] Acquiring lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.989865] env[61824]: DEBUG oslo_concurrency.lockutils [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.989998] env[61824]: DEBUG oslo_concurrency.lockutils [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.990193] env[61824]: DEBUG nova.compute.manager [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] No waiting events found dispatching network-vif-plugged-a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 897.990371] env[61824]: WARNING nova.compute.manager [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received unexpected event network-vif-plugged-a608c7e2-51a6-4791-b00b-570fb8a353b0 for instance with vm_state active and task_state None. [ 897.990586] env[61824]: DEBUG nova.compute.manager [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-changed-a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 897.990765] env[61824]: DEBUG nova.compute.manager [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing instance network info cache due to event network-changed-a608c7e2-51a6-4791-b00b-570fb8a353b0. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 897.990941] env[61824]: DEBUG oslo_concurrency.lockutils [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.999753] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52244572-a2c7-79e1-55ee-edc00ca30a77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.021672] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updated VIF entry in instance network info cache for port bd35af69-df49-4c0a-a798-57bd9c8f6fe4. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 898.022097] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.070611] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.070730] env[61824]: DEBUG nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.073292] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.974s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.073515] env[61824]: DEBUG nova.objects.instance [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lazy-loading 'resources' on Instance uuid 01fd9faa-a0de-45b2-89bf-094c6d1ac155 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.183854] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.444076] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275762, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.499590] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52244572-a2c7-79e1-55ee-edc00ca30a77, 'name': SearchDatastore_Task, 'duration_secs': 0.014542} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.499961] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.500297] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.500686] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.500820] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.501075] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.501428] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41792478-a490-48f6-a2a9-e84b096724ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.513721] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.513956] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.514801] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ead23a64-f047-4869-a760-885e41b32458 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.520320] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 898.520320] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d72657-acf4-3370-b8c6-c77dece05809" [ 898.520320] env[61824]: _type = "Task" [ 898.520320] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.525245] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.525506] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Received event network-vif-plugged-41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.525700] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Acquiring lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.525905] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.526119] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.526331] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] No waiting events found dispatching network-vif-plugged-41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 898.526502] env[61824]: WARNING nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Received unexpected event network-vif-plugged-41458d47-a6e3-44c2-a378-9968713e2739 for instance with vm_state building and task_state spawning. [ 898.526673] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Received event network-changed-41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.526828] env[61824]: DEBUG nova.compute.manager [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Refreshing instance network info cache due to event network-changed-41458d47-a6e3-44c2-a378-9968713e2739. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.527036] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Acquiring lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.527184] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Acquired lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.527351] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Refreshing network info cache for port 41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.532066] env[61824]: DEBUG oslo_concurrency.lockutils [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.532066] env[61824]: DEBUG nova.network.neutron [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Refreshing network info cache for port a608c7e2-51a6-4791-b00b-570fb8a353b0 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.532599] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d72657-acf4-3370-b8c6-c77dece05809, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.576324] env[61824]: DEBUG nova.compute.utils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.579555] env[61824]: DEBUG nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.579555] env[61824]: DEBUG nova.network.neutron [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 898.667448] env[61824]: DEBUG nova.policy [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f1d7f006ab244da48e1bfc38a5859932', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e23c7b193932478096b2fbf49a59fd7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.828261] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aeff516-2fa0-4c41-9898-cf4caf91d704 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.838310] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b2d05b-c554-4e95-bde2-feadade0e56b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.879055] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1180933f-2ceb-4c5a-a08b-961f57587674 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.887854] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1d5679-96a0-46ae-bab2-7e5baf96cd6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.902289] env[61824]: DEBUG nova.compute.provider_tree [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.944390] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275762, 'name': ReconfigVM_Task, 'duration_secs': 0.7635} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.944684] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5/07973a77-0fe8-4bc9-b8c5-e693d7a63ef5.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.945373] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5024039b-8b86-43a5-99ec-abfb8cd0fcbb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.952256] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 898.952256] env[61824]: value = "task-1275763" [ 898.952256] env[61824]: _type = "Task" [ 898.952256] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.961120] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275763, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.031182] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d72657-acf4-3370-b8c6-c77dece05809, 'name': SearchDatastore_Task, 'duration_secs': 0.019555} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.032148] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e76e3013-a086-49db-abab-98b5f456affc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.040661] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 899.040661] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52647cbe-fe7a-912b-df02-f50c74646792" [ 899.040661] env[61824]: _type = "Task" [ 899.040661] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.048958] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52647cbe-fe7a-912b-df02-f50c74646792, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.080018] env[61824]: DEBUG nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.083601] env[61824]: DEBUG nova.network.neutron [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Successfully created port: 12e4609d-03ba-48c6-9526-a2c8e2c21b90 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.407198] env[61824]: DEBUG nova.scheduler.client.report [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.471868] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275763, 'name': Rename_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.551568] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52647cbe-fe7a-912b-df02-f50c74646792, 'name': SearchDatastore_Task, 'duration_secs': 0.012783} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.551727] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.552193] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8/0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.552318] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43d377af-404a-48ad-8716-36a937205954 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.559416] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 899.559416] env[61824]: value = "task-1275764" [ 899.559416] env[61824]: _type = "Task" [ 899.559416] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.567547] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275764, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.600265] env[61824]: DEBUG nova.network.neutron [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Added VIF to instance network info cache for port a608c7e2-51a6-4791-b00b-570fb8a353b0. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3489}} [ 899.600448] env[61824]: DEBUG nova.network.neutron [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "address": "fa:16:3e:da:35:5b", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa608c7e2-51", "ovs_interfaceid": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.607739] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Updated VIF entry in instance network info cache for port 41458d47-a6e3-44c2-a378-9968713e2739. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 899.608818] env[61824]: DEBUG nova.network.neutron [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Updating instance_info_cache with network_info: [{"id": "41458d47-a6e3-44c2-a378-9968713e2739", "address": "fa:16:3e:96:67:50", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41458d47-a6", "ovs_interfaceid": "41458d47-a6e3-44c2-a378-9968713e2739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.911922] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.915511] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.814s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.915889] env[61824]: DEBUG nova.objects.instance [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lazy-loading 'resources' on Instance uuid 2495fcb4-28cb-49c3-90fe-c84072466287 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.949744] env[61824]: INFO nova.scheduler.client.report [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted allocations for instance 01fd9faa-a0de-45b2-89bf-094c6d1ac155 [ 899.970084] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275763, 'name': Rename_Task, 'duration_secs': 0.804159} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.970411] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.970675] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55ac337d-c4e2-4272-86c9-a9d56372fe3b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.979411] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 899.979411] env[61824]: value = "task-1275765" [ 899.979411] env[61824]: _type = "Task" [ 899.979411] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.989126] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.071303] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275764, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.096053] env[61824]: DEBUG nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.103647] env[61824]: DEBUG oslo_concurrency.lockutils [req-08bccf1e-2821-49dd-a413-78e3f408b516 req-b589f3ea-6beb-4724-9517-0c5b771feb72 service nova] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.103838] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.104036] env[61824]: DEBUG nova.network.neutron [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 900.110680] env[61824]: DEBUG oslo_concurrency.lockutils [req-67dc0ac4-6dc2-4574-9deb-408fb05d1817 req-d74f8dab-e28e-44ea-99f7-b6cafc0b1910 service nova] Releasing lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.125419] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.125554] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.125627] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.125811] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.125957] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.126138] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.126506] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.126569] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.126735] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.126901] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.127167] env[61824]: DEBUG nova.virt.hardware [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.128221] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9e915b-a317-4794-8c20-74f6c320a838 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.136351] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce62019d-d7d2-49f4-be38-8cc6d6451ba4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.466710] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d6126cff-0afd-4695-a315-7abeb101eb3d tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "01fd9faa-a0de-45b2-89bf-094c6d1ac155" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.226s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.490972] env[61824]: DEBUG oslo_vmware.api [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275765, 'name': PowerOnVM_Task, 'duration_secs': 0.461388} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.492166] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.492386] env[61824]: INFO nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Took 8.83 seconds to spawn the instance on the hypervisor. [ 900.492573] env[61824]: DEBUG nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.493763] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa52824-52c2-42af-87be-32f6ec72c50a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.571724] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275764, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53161} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.574275] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8/0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 900.574502] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.574931] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96f4f7c0-1e66-4c64-99e7-bd2695a8806b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.581181] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 900.581181] env[61824]: value = "task-1275766" [ 900.581181] env[61824]: _type = "Task" [ 900.581181] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.591458] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.635029] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6154f562-325c-473a-a336-f3a875a38ae6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.643939] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9669b85f-f934-43e8-8a05-b34c4f57f810 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.653263] env[61824]: WARNING nova.network.neutron [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] 34841295-0562-4caa-8275-c3695a63d719 already exists in list: networks containing: ['34841295-0562-4caa-8275-c3695a63d719']. ignoring it [ 900.654025] env[61824]: WARNING nova.network.neutron [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] 34841295-0562-4caa-8275-c3695a63d719 already exists in list: networks containing: ['34841295-0562-4caa-8275-c3695a63d719']. ignoring it [ 900.654025] env[61824]: WARNING nova.network.neutron [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] a608c7e2-51a6-4791-b00b-570fb8a353b0 already exists in list: port_ids containing: ['a608c7e2-51a6-4791-b00b-570fb8a353b0']. ignoring it [ 900.689324] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-135ec87e-b6dc-40cc-af63-8560652960fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.700203] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef2e769-e6d4-4f3e-bd40-86423532ccaa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.717591] env[61824]: DEBUG nova.compute.provider_tree [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.736515] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "9f031313-766b-4696-9bc6-75ced078bfde" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.736809] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "9f031313-766b-4696-9bc6-75ced078bfde" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.737044] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "9f031313-766b-4696-9bc6-75ced078bfde-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.737251] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "9f031313-766b-4696-9bc6-75ced078bfde-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.737458] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "9f031313-766b-4696-9bc6-75ced078bfde-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.741766] env[61824]: INFO nova.compute.manager [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Terminating instance [ 900.745485] env[61824]: DEBUG nova.compute.manager [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 900.749328] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.749328] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66bbe72-3efd-4c79-a127-b017148227af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.756978] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.757086] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2ad9fef-dd84-40d0-80c9-5624a74479ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.764312] env[61824]: DEBUG oslo_vmware.api [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 900.764312] env[61824]: value = "task-1275767" [ 900.764312] env[61824]: _type = "Task" [ 900.764312] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.774137] env[61824]: DEBUG oslo_vmware.api [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275767, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.011700] env[61824]: INFO nova.compute.manager [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Took 19.52 seconds to build instance. [ 901.091266] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070445} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.091560] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.092347] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4166e18-2966-4200-90b5-f5486c4befc2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.116142] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8/0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.116480] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33e9f1fb-8b63-497d-a4d5-5a9388b71e02 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.141395] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 901.141395] env[61824]: value = "task-1275768" [ 901.141395] env[61824]: _type = "Task" [ 901.141395] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.152682] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275768, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.247154] env[61824]: ERROR nova.scheduler.client.report [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [req-1e8fb812-3e29-48e5-b5bb-fae23fa70cdd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8d9ece0e-729c-4d69-b741-d89969200860. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1e8fb812-3e29-48e5-b5bb-fae23fa70cdd"}]} [ 901.263779] env[61824]: DEBUG nova.scheduler.client.report [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing inventories for resource provider 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 901.276609] env[61824]: DEBUG oslo_vmware.api [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275767, 'name': PowerOffVM_Task, 'duration_secs': 0.195263} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.276909] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.277090] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.277346] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7726012a-c2f8-407c-8aeb-312866749e45 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.280437] env[61824]: DEBUG nova.scheduler.client.report [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating ProviderTree inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 901.280702] env[61824]: DEBUG nova.compute.provider_tree [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.292350] env[61824]: DEBUG nova.scheduler.client.report [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing aggregate associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, aggregates: None {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 901.317846] env[61824]: DEBUG nova.scheduler.client.report [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Refreshing trait associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 901.338441] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.338661] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.338987] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Deleting the datastore file [datastore1] 9f031313-766b-4696-9bc6-75ced078bfde {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.339094] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a90edec-2ec4-40ba-9f17-ccadde92fc9e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.345666] env[61824]: DEBUG oslo_vmware.api [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for the task: (returnval){ [ 901.345666] env[61824]: value = "task-1275770" [ 901.345666] env[61824]: _type = "Task" [ 901.345666] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.353737] env[61824]: DEBUG oslo_vmware.api [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.471821] env[61824]: DEBUG nova.network.neutron [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "address": "fa:16:3e:da:35:5b", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa608c7e2-51", "ovs_interfaceid": "a608c7e2-51a6-4791-b00b-570fb8a353b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.517953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a83957aa-23d1-431c-b0bd-dc3a44c1264f tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.043s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.543394] env[61824]: DEBUG nova.network.neutron [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Successfully updated port: 12e4609d-03ba-48c6-9526-a2c8e2c21b90 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 901.585628] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "e5825315-bfea-4eec-80ca-cb68d9748d22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.586036] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.623301] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44caa757-e1e3-4e1d-95cd-2715cd452880 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.630835] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.631212] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.636999] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e114b3c3-dba0-4e4e-8a9e-2259ea38d376 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.677412] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764bbc3e-8efb-482d-8fa9-8ee7078fdba4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.681395] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.681633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.686603] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275768, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.690831] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a6b592-b9d7-4de7-8315-281a341aac3e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.705797] env[61824]: DEBUG nova.compute.provider_tree [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.857140] env[61824]: DEBUG oslo_vmware.api [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Task: {'id': task-1275770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.496291} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.858206] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.858206] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.858206] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.858206] env[61824]: INFO nova.compute.manager [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Took 1.11 seconds to destroy the instance on the hypervisor. [ 901.858907] env[61824]: DEBUG oslo.service.loopingcall [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.858907] env[61824]: DEBUG nova.compute.manager [-] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.858907] env[61824]: DEBUG nova.network.neutron [-] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 901.974510] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.975440] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.975440] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.976273] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cc9ce9-02fb-47ec-a510-34efd7767d51 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.995814] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 901.996099] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 901.996255] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 901.996514] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 901.996579] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 901.996737] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 901.996937] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 901.997202] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 901.997287] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 901.997445] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 901.997634] env[61824]: DEBUG nova.virt.hardware [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.003879] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Reconfiguring VM to attach interface {{(pid=61824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 902.004219] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9a7ccee-1f9a-462f-98b8-89bbf44d02fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.021709] env[61824]: DEBUG oslo_vmware.api [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 902.021709] env[61824]: value = "task-1275771" [ 902.021709] env[61824]: _type = "Task" [ 902.021709] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.027021] env[61824]: DEBUG nova.compute.manager [req-c4c32aa8-9216-4903-8cef-f87d7061256e req-43aa24ac-3818-4fe3-b27d-ac25b39a96a2 service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Received event network-vif-plugged-12e4609d-03ba-48c6-9526-a2c8e2c21b90 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.027245] env[61824]: DEBUG oslo_concurrency.lockutils [req-c4c32aa8-9216-4903-8cef-f87d7061256e req-43aa24ac-3818-4fe3-b27d-ac25b39a96a2 service nova] Acquiring lock "beafe3e0-ff41-40f7-90cf-1d7977191180-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.027454] env[61824]: DEBUG oslo_concurrency.lockutils [req-c4c32aa8-9216-4903-8cef-f87d7061256e req-43aa24ac-3818-4fe3-b27d-ac25b39a96a2 service nova] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.027621] env[61824]: DEBUG oslo_concurrency.lockutils [req-c4c32aa8-9216-4903-8cef-f87d7061256e req-43aa24ac-3818-4fe3-b27d-ac25b39a96a2 service nova] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.027845] env[61824]: DEBUG nova.compute.manager [req-c4c32aa8-9216-4903-8cef-f87d7061256e req-43aa24ac-3818-4fe3-b27d-ac25b39a96a2 service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] No waiting events found dispatching network-vif-plugged-12e4609d-03ba-48c6-9526-a2c8e2c21b90 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.028167] env[61824]: WARNING nova.compute.manager [req-c4c32aa8-9216-4903-8cef-f87d7061256e req-43aa24ac-3818-4fe3-b27d-ac25b39a96a2 service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Received unexpected event network-vif-plugged-12e4609d-03ba-48c6-9526-a2c8e2c21b90 for instance with vm_state building and task_state spawning. [ 902.033672] env[61824]: DEBUG oslo_vmware.api [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275771, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.045957] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "refresh_cache-beafe3e0-ff41-40f7-90cf-1d7977191180" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.045957] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "refresh_cache-beafe3e0-ff41-40f7-90cf-1d7977191180" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.045957] env[61824]: DEBUG nova.network.neutron [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 902.092022] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.135768] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.152151] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275768, 'name': ReconfigVM_Task, 'duration_secs': 0.793561} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.152698] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8/0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.153526] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9f079d44-8143-4b6e-b842-ddb4e642162e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.162846] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 902.162846] env[61824]: value = "task-1275772" [ 902.162846] env[61824]: _type = "Task" [ 902.162846] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.173424] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275772, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.189016] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.246019] env[61824]: DEBUG nova.scheduler.client.report [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updated inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 902.246019] env[61824]: DEBUG nova.compute.provider_tree [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating resource provider 8d9ece0e-729c-4d69-b741-d89969200860 generation from 103 to 104 during operation: update_inventory {{(pid=61824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 902.246019] env[61824]: DEBUG nova.compute.provider_tree [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 141, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 902.537176] env[61824]: DEBUG oslo_vmware.api [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275771, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.587174] env[61824]: DEBUG nova.network.neutron [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 902.614932] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.662533] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.668310] env[61824]: DEBUG nova.network.neutron [-] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.673715] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275772, 'name': Rename_Task, 'duration_secs': 0.210476} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.674430] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 902.674555] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e131665-87bf-49b3-bfe4-0011c4106720 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.683082] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 902.683082] env[61824]: value = "task-1275773" [ 902.683082] env[61824]: _type = "Task" [ 902.683082] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.695396] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.710575] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.750612] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.835s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.752753] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 15.939s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.775016] env[61824]: INFO nova.scheduler.client.report [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted allocations for instance 2495fcb4-28cb-49c3-90fe-c84072466287 [ 902.795505] env[61824]: DEBUG nova.network.neutron [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Updating instance_info_cache with network_info: [{"id": "12e4609d-03ba-48c6-9526-a2c8e2c21b90", "address": "fa:16:3e:64:a7:21", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12e4609d-03", "ovs_interfaceid": "12e4609d-03ba-48c6-9526-a2c8e2c21b90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.033120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.033387] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.033606] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.033804] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.033985] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.035658] env[61824]: DEBUG oslo_vmware.api [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275771, 'name': ReconfigVM_Task, 'duration_secs': 0.547453} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.036114] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.036439] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Reconfigured VM to attach interface {{(pid=61824) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 903.039199] env[61824]: INFO nova.compute.manager [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Terminating instance [ 903.041692] env[61824]: DEBUG nova.compute.manager [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 903.041883] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.042753] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a639aa-3fb9-4559-be80-b49fd9da9f7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.050236] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.050535] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-912ffbec-b303-4b1f-81f0-73f48c5cf96e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.057599] env[61824]: DEBUG oslo_vmware.api [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 903.057599] env[61824]: value = "task-1275774" [ 903.057599] env[61824]: _type = "Task" [ 903.057599] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.065521] env[61824]: DEBUG oslo_vmware.api [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275774, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.175376] env[61824]: INFO nova.compute.manager [-] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Took 1.32 seconds to deallocate network for instance. [ 903.194523] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275773, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.258322] env[61824]: INFO nova.compute.claims [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.285052] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bccea415-9aaf-404a-a359-62a25d27fc9d tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "2495fcb4-28cb-49c3-90fe-c84072466287" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.115s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.299547] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "refresh_cache-beafe3e0-ff41-40f7-90cf-1d7977191180" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.299698] env[61824]: DEBUG nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Instance network_info: |[{"id": "12e4609d-03ba-48c6-9526-a2c8e2c21b90", "address": "fa:16:3e:64:a7:21", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12e4609d-03", "ovs_interfaceid": "12e4609d-03ba-48c6-9526-a2c8e2c21b90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 903.300719] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:64:a7:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7f41333-42ee-47f3-936c-d6701ab786d2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12e4609d-03ba-48c6-9526-a2c8e2c21b90', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 903.307566] env[61824]: DEBUG oslo.service.loopingcall [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.308053] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 903.308317] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52412701-81b0-4954-837b-aa6d1e9518c9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.329007] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 903.329007] env[61824]: value = "task-1275775" [ 903.329007] env[61824]: _type = "Task" [ 903.329007] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.336885] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275775, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.342710] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.342960] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.343192] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.343383] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.343599] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.345682] env[61824]: INFO nova.compute.manager [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Terminating instance [ 903.347834] env[61824]: DEBUG nova.compute.manager [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 903.347834] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.348595] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806b13a0-318c-41ff-ad63-8cf60360f20d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.357816] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.358105] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9872c022-e1bd-4490-9313-6c5bfebbe849 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.364967] env[61824]: DEBUG oslo_vmware.api [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 903.364967] env[61824]: value = "task-1275776" [ 903.364967] env[61824]: _type = "Task" [ 903.364967] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.371424] env[61824]: DEBUG oslo_vmware.api [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.542448] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7b12a4a5-3e31-44d7-ab7a-3ceb04513785 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-aa9c9eb2-b411-4eca-9f8d-3df377ffe351-a608c7e2-51a6-4791-b00b-570fb8a353b0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.881s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.568493] env[61824]: DEBUG oslo_vmware.api [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275774, 'name': PowerOffVM_Task, 'duration_secs': 0.205239} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.568787] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.568959] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.569231] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51f3c5aa-0cab-45dd-81c5-67637f8d80b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.623838] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.624092] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.624284] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleting the datastore file [datastore1] d6576f45-35f6-48dd-9160-80fc53d8b6b2 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.624554] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-afbffc01-5c46-47e7-9aef-a62675d05165 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.631093] env[61824]: DEBUG oslo_vmware.api [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 903.631093] env[61824]: value = "task-1275778" [ 903.631093] env[61824]: _type = "Task" [ 903.631093] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.649126] env[61824]: DEBUG oslo_vmware.api [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.684433] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.693672] env[61824]: DEBUG oslo_vmware.api [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275773, 'name': PowerOnVM_Task, 'duration_secs': 0.585097} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.693939] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 903.694202] env[61824]: INFO nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Took 9.64 seconds to spawn the instance on the hypervisor. [ 903.694419] env[61824]: DEBUG nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 903.695201] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6b4f68-7d2b-44ba-a926-419b5197cf71 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.768067] env[61824]: INFO nova.compute.resource_tracker [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating resource usage from migration 208ec0f4-8708-472d-b4e9-bcd663ed20d5 [ 903.842753] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275775, 'name': CreateVM_Task, 'duration_secs': 0.386814} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.843176] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 903.843926] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.844140] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.844514] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.844801] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92759877-e883-4ed9-9ed3-c8b9c2253fed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.849851] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 903.849851] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52fb69a3-8c2e-4e2e-45e6-a21460309235" [ 903.849851] env[61824]: _type = "Task" [ 903.849851] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.860859] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fb69a3-8c2e-4e2e-45e6-a21460309235, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.872338] env[61824]: DEBUG oslo_vmware.api [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275776, 'name': PowerOffVM_Task, 'duration_secs': 0.242942} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.875282] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.875405] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.875848] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25a1becb-a6a8-46f8-a24f-00cb380fab43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.961819] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.962132] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.962976] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleting the datastore file [datastore1] 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.962976] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7dd3009-91e6-4c58-b55e-321ad469cc79 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.968233] env[61824]: DEBUG oslo_vmware.api [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 903.968233] env[61824]: value = "task-1275780" [ 903.968233] env[61824]: _type = "Task" [ 903.968233] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.979091] env[61824]: DEBUG oslo_vmware.api [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.017678] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0a0f48-f268-48c8-9c82-21a1ea7678a4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.025340] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562d7080-eabe-430b-9ed2-23c6eba1c4df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.057931] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751f13e5-2b42-4b31-a9e2-b527b6d12fb9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.065658] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da58f429-0f2e-41a2-bb10-0c2c202d3dba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.081049] env[61824]: DEBUG nova.compute.provider_tree [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.140403] env[61824]: DEBUG oslo_vmware.api [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26402} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.140689] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.140879] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.141067] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.141246] env[61824]: INFO nova.compute.manager [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Took 1.10 seconds to destroy the instance on the hypervisor. [ 904.141688] env[61824]: DEBUG oslo.service.loopingcall [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.141688] env[61824]: DEBUG nova.compute.manager [-] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.141911] env[61824]: DEBUG nova.network.neutron [-] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 904.167035] env[61824]: DEBUG nova.compute.manager [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Received event network-changed-12e4609d-03ba-48c6-9526-a2c8e2c21b90 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.167246] env[61824]: DEBUG nova.compute.manager [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Refreshing instance network info cache due to event network-changed-12e4609d-03ba-48c6-9526-a2c8e2c21b90. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.167476] env[61824]: DEBUG oslo_concurrency.lockutils [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] Acquiring lock "refresh_cache-beafe3e0-ff41-40f7-90cf-1d7977191180" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.167627] env[61824]: DEBUG oslo_concurrency.lockutils [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] Acquired lock "refresh_cache-beafe3e0-ff41-40f7-90cf-1d7977191180" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.167788] env[61824]: DEBUG nova.network.neutron [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Refreshing network info cache for port 12e4609d-03ba-48c6-9526-a2c8e2c21b90 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.211124] env[61824]: INFO nova.compute.manager [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Took 22.31 seconds to build instance. [ 904.372098] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fb69a3-8c2e-4e2e-45e6-a21460309235, 'name': SearchDatastore_Task, 'duration_secs': 0.009858} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.372435] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.372679] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.372919] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.373085] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.373272] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.373557] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c52fb1d7-cff0-486a-910e-225436eebc21 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.387436] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.387436] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 904.387436] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2d51697-b4e6-4ef1-b6db-921191934e52 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.394490] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 904.394490] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5234930e-57af-d46e-0cae-5d747b1f0089" [ 904.394490] env[61824]: _type = "Task" [ 904.394490] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.401622] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5234930e-57af-d46e-0cae-5d747b1f0089, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.478463] env[61824]: DEBUG oslo_vmware.api [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.622564] env[61824]: DEBUG nova.scheduler.client.report [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Updated inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 with generation 104 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 904.627359] env[61824]: DEBUG nova.compute.provider_tree [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Updating resource provider 8d9ece0e-729c-4d69-b741-d89969200860 generation from 104 to 105 during operation: update_inventory {{(pid=61824) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 904.627359] env[61824]: DEBUG nova.compute.provider_tree [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 904.720632] env[61824]: DEBUG oslo_concurrency.lockutils [None req-548b31bc-f5b2-4ad8-a7c4-c29e8c819ed1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.828s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.904237] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5234930e-57af-d46e-0cae-5d747b1f0089, 'name': SearchDatastore_Task, 'duration_secs': 0.011127} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.905049] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-792f0d62-ca6a-447a-9cf1-f6e5915754ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.910018] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 904.910018] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52536953-1c41-2222-e1bd-55a4d20854dd" [ 904.910018] env[61824]: _type = "Task" [ 904.910018] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.919779] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52536953-1c41-2222-e1bd-55a4d20854dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.980403] env[61824]: DEBUG oslo_vmware.api [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.600904} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.980719] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.980852] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.981110] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.981223] env[61824]: INFO nova.compute.manager [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Took 1.63 seconds to destroy the instance on the hypervisor. [ 904.981464] env[61824]: DEBUG oslo.service.loopingcall [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.981717] env[61824]: DEBUG nova.compute.manager [-] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.981777] env[61824]: DEBUG nova.network.neutron [-] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 905.058387] env[61824]: DEBUG nova.network.neutron [-] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.133636] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.381s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.133860] env[61824]: INFO nova.compute.manager [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Migrating [ 905.140762] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.363s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.140976] env[61824]: DEBUG nova.objects.instance [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'pci_requests' on Instance uuid 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.151176] env[61824]: DEBUG nova.objects.instance [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'numa_topology' on Instance uuid 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.158774] env[61824]: DEBUG nova.network.neutron [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Updated VIF entry in instance network info cache for port 12e4609d-03ba-48c6-9526-a2c8e2c21b90. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.158774] env[61824]: DEBUG nova.network.neutron [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Updating instance_info_cache with network_info: [{"id": "12e4609d-03ba-48c6-9526-a2c8e2c21b90", "address": "fa:16:3e:64:a7:21", "network": {"id": "eecde403-cec7-4a2f-ac0a-313776b7fbcc", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1923184865-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e23c7b193932478096b2fbf49a59fd7e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12e4609d-03", "ovs_interfaceid": "12e4609d-03ba-48c6-9526-a2c8e2c21b90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.315278] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "interface-aa9c9eb2-b411-4eca-9f8d-3df377ffe351-a608c7e2-51a6-4791-b00b-570fb8a353b0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.315573] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-aa9c9eb2-b411-4eca-9f8d-3df377ffe351-a608c7e2-51a6-4791-b00b-570fb8a353b0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.422028] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52536953-1c41-2222-e1bd-55a4d20854dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009352} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.422156] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.422431] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] beafe3e0-ff41-40f7-90cf-1d7977191180/beafe3e0-ff41-40f7-90cf-1d7977191180.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 905.422712] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f292e44d-9efc-4524-b290-975a3980a7ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.429493] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 905.429493] env[61824]: value = "task-1275781" [ 905.429493] env[61824]: _type = "Task" [ 905.429493] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.439441] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275781, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.488396] env[61824]: DEBUG nova.compute.manager [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Received event network-changed-41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.488646] env[61824]: DEBUG nova.compute.manager [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Refreshing instance network info cache due to event network-changed-41458d47-a6e3-44c2-a378-9968713e2739. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.488870] env[61824]: DEBUG oslo_concurrency.lockutils [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] Acquiring lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.489018] env[61824]: DEBUG oslo_concurrency.lockutils [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] Acquired lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.489313] env[61824]: DEBUG nova.network.neutron [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Refreshing network info cache for port 41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.560970] env[61824]: INFO nova.compute.manager [-] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Took 1.42 seconds to deallocate network for instance. [ 905.652980] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.653334] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.653372] env[61824]: DEBUG nova.network.neutron [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 905.656231] env[61824]: INFO nova.compute.claims [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.664012] env[61824]: DEBUG oslo_concurrency.lockutils [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] Releasing lock "refresh_cache-beafe3e0-ff41-40f7-90cf-1d7977191180" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.664012] env[61824]: DEBUG nova.compute.manager [req-a44f3d17-e87c-4fa1-96c4-7c0075b30a61 req-c8eb27d0-f42c-4f2b-9a6d-5e651d3efc6f service nova] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Received event network-vif-deleted-adefa466-eb42-422d-ade5-3f7cf6e305ef {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.819171] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.819610] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.820374] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c87281f-ed01-47a0-8a6b-97478d42babd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.849953] env[61824]: DEBUG nova.network.neutron [-] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.855424] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6a6da7-5fc9-407b-9e78-6a6d52febf64 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.885680] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Reconfiguring VM to detach interface {{(pid=61824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 905.886757] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b4b555d-225a-4519-8f9d-425e4a673d5b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.905948] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 905.905948] env[61824]: value = "task-1275782" [ 905.905948] env[61824]: _type = "Task" [ 905.905948] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.914192] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.940569] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275781, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445845} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.940872] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] beafe3e0-ff41-40f7-90cf-1d7977191180/beafe3e0-ff41-40f7-90cf-1d7977191180.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 905.941140] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.941380] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15094a43-cbd6-40bd-8f26-527814840c71 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.947899] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 905.947899] env[61824]: value = "task-1275783" [ 905.947899] env[61824]: _type = "Task" [ 905.947899] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.956379] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275783, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.068080] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.355461] env[61824]: INFO nova.compute.manager [-] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Took 1.37 seconds to deallocate network for instance. [ 906.416533] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.458360] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275783, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082355} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.458750] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.459511] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8fdbc9-70fa-4e76-aadb-89d9a193e60b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.482265] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] beafe3e0-ff41-40f7-90cf-1d7977191180/beafe3e0-ff41-40f7-90cf-1d7977191180.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.482540] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18c9f31f-9237-46f2-a332-b22182e75ee6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.508956] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 906.508956] env[61824]: value = "task-1275784" [ 906.508956] env[61824]: _type = "Task" [ 906.508956] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.518607] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275784, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.531353] env[61824]: DEBUG nova.network.neutron [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Updated VIF entry in instance network info cache for port 41458d47-a6e3-44c2-a378-9968713e2739. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.531868] env[61824]: DEBUG nova.network.neutron [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Updating instance_info_cache with network_info: [{"id": "41458d47-a6e3-44c2-a378-9968713e2739", "address": "fa:16:3e:96:67:50", "network": {"id": "a9540aa2-1f80-4d47-af16-269cd59f7796", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1020087775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.135", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "502b41859ba843838d2188c82f70ae08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap41458d47-a6", "ovs_interfaceid": "41458d47-a6e3-44c2-a378-9968713e2739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.550329] env[61824]: DEBUG nova.compute.manager [req-99fc21d4-659e-4220-8362-0a98e40719a8 req-4199dd24-f23f-4c4e-880a-acc530201f8c service nova] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Received event network-vif-deleted-3734404d-9e67-4190-ab94-d0fd50c732c6 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.550595] env[61824]: DEBUG nova.compute.manager [req-99fc21d4-659e-4220-8362-0a98e40719a8 req-4199dd24-f23f-4c4e-880a-acc530201f8c service nova] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Received event network-vif-deleted-734e041a-690b-4a80-9b58-6c6bcf162752 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.567351] env[61824]: DEBUG nova.network.neutron [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance_info_cache with network_info: [{"id": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "address": "fa:16:3e:ad:9e:74", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap542a267a-6d", "ovs_interfaceid": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.865779] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.889753] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997ba551-139d-44b5-ac79-11e7dbc3e7af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.898869] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61cb92a-e19a-423e-bd5c-83990950fbe9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.930022] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef06d8d-6e60-4b35-8a97-0800733a124e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.937530] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.940796] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184ad5e3-c969-4500-9898-985ae09f747f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.953711] env[61824]: DEBUG nova.compute.provider_tree [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.020889] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275784, 'name': ReconfigVM_Task, 'duration_secs': 0.256647} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.020889] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Reconfigured VM instance instance-00000054 to attach disk [datastore2] beafe3e0-ff41-40f7-90cf-1d7977191180/beafe3e0-ff41-40f7-90cf-1d7977191180.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 907.021275] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07745eba-35a1-450b-893a-ca22a43f9964 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.028261] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 907.028261] env[61824]: value = "task-1275785" [ 907.028261] env[61824]: _type = "Task" [ 907.028261] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.036642] env[61824]: DEBUG oslo_concurrency.lockutils [req-56936c27-281e-441b-8b4e-4f1d5a6b74a8 req-b9f96589-9b16-49ee-be08-cb1f3fdf260e service nova] Releasing lock "refresh_cache-0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.037067] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275785, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.070192] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.436265] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.456404] env[61824]: DEBUG nova.scheduler.client.report [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.538094] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275785, 'name': Rename_Task, 'duration_secs': 0.128046} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.538420] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.538668] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ac889294-09c5-4c2d-9849-841ff5ff2131 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.543979] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 907.543979] env[61824]: value = "task-1275786" [ 907.543979] env[61824]: _type = "Task" [ 907.543979] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.551032] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275786, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.937512] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.961472] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.821s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.963700] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.696s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.963956] env[61824]: DEBUG nova.objects.instance [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lazy-loading 'resources' on Instance uuid cbded637-56b9-4048-b66e-1edf6311f435 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.026700] env[61824]: INFO nova.network.neutron [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating port a66cb8c8-5826-4e02-ad96-b2bbea85d23b with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 908.054236] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275786, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.439000] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.557101] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275786, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.588632] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7010939-1925-4a83-a3ca-9ab2499e936b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.613510] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance '09432e72-c10a-4a7a-93af-af0d371fb871' progress to 0 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 908.726145] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c347625-9b7a-415e-a8e5-fb459497c730 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.735262] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fda17e-da37-4157-82c1-7d0b4ee37007 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.764907] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3dca55-3809-4bb1-85f9-f291cedb1cfc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.772614] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7280c3e-2f0b-4cae-9160-10a6412d8908 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.786292] env[61824]: DEBUG nova.compute.provider_tree [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.940098] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.056243] env[61824]: DEBUG oslo_vmware.api [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275786, 'name': PowerOnVM_Task, 'duration_secs': 1.443691} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.056539] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 909.056746] env[61824]: INFO nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Took 8.96 seconds to spawn the instance on the hypervisor. [ 909.056928] env[61824]: DEBUG nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 909.057734] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3804604-270e-40ed-8b07-c3ed230a0c09 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.122211] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.122522] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19854458-f3ac-45be-a042-c3d992a842d9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.130143] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 909.130143] env[61824]: value = "task-1275787" [ 909.130143] env[61824]: _type = "Task" [ 909.130143] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.137896] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275787, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.290056] env[61824]: DEBUG nova.scheduler.client.report [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.441647] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.575021] env[61824]: INFO nova.compute.manager [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Took 25.63 seconds to build instance. [ 909.642124] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275787, 'name': PowerOffVM_Task, 'duration_secs': 0.186703} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.642447] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.642916] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance '09432e72-c10a-4a7a-93af-af0d371fb871' progress to 17 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 909.794072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.830s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.796814] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.182s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.799614] env[61824]: INFO nova.compute.claims [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 909.815525] env[61824]: INFO nova.scheduler.client.report [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Deleted allocations for instance cbded637-56b9-4048-b66e-1edf6311f435 [ 909.943142] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.077323] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b939434d-d591-4d18-bc3d-93a0a253f64f tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.148s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.080610] env[61824]: DEBUG nova.compute.manager [req-2d27fe68-6b47-44bc-9152-f3349b6f1119 req-5d65fd30-637d-4f53-820c-1e8572a9a90d service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-vif-plugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.080830] env[61824]: DEBUG oslo_concurrency.lockutils [req-2d27fe68-6b47-44bc-9152-f3349b6f1119 req-5d65fd30-637d-4f53-820c-1e8572a9a90d service nova] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.081429] env[61824]: DEBUG oslo_concurrency.lockutils [req-2d27fe68-6b47-44bc-9152-f3349b6f1119 req-5d65fd30-637d-4f53-820c-1e8572a9a90d service nova] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.081621] env[61824]: DEBUG oslo_concurrency.lockutils [req-2d27fe68-6b47-44bc-9152-f3349b6f1119 req-5d65fd30-637d-4f53-820c-1e8572a9a90d service nova] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.081803] env[61824]: DEBUG nova.compute.manager [req-2d27fe68-6b47-44bc-9152-f3349b6f1119 req-5d65fd30-637d-4f53-820c-1e8572a9a90d service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] No waiting events found dispatching network-vif-plugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.081976] env[61824]: WARNING nova.compute.manager [req-2d27fe68-6b47-44bc-9152-f3349b6f1119 req-5d65fd30-637d-4f53-820c-1e8572a9a90d service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received unexpected event network-vif-plugged-a66cb8c8-5826-4e02-ad96-b2bbea85d23b for instance with vm_state shelved_offloaded and task_state spawning. [ 910.105620] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.105742] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.105896] env[61824]: DEBUG nova.network.neutron [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 910.151448] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.151448] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.151667] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.151899] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.152174] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.152445] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.152798] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.153107] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.153396] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.153680] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.153974] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.163443] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa5ee19f-f8da-44fb-a3e8-8c559fdd86fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.189389] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 910.189389] env[61824]: value = "task-1275788" [ 910.189389] env[61824]: _type = "Task" [ 910.189389] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.202782] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275788, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.325297] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8228b339-8a9c-43fe-9caa-e8c3efdad6c4 tempest-AttachVolumeTestJSON-1146451136 tempest-AttachVolumeTestJSON-1146451136-project-member] Lock "cbded637-56b9-4048-b66e-1edf6311f435" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.760s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.442057] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.700732] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275788, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.944165] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.007063] env[61824]: DEBUG nova.compute.manager [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.008006] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d808395-b636-43b3-8452-601c57ff996f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.095942] env[61824]: DEBUG nova.network.neutron [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.131766] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e82c84-e046-4a10-a9fc-6c1928fc5eed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.140516] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8f6409-b30b-40d8-8546-94ee594c1904 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.172502] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa1da68-3ad5-42a5-bf8b-b5f1de0f9d93 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.181058] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215eba47-57bb-4f43-ac0b-2d0898b2bc22 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.196384] env[61824]: DEBUG nova.compute.provider_tree [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.205255] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275788, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.444511] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.529680] env[61824]: INFO nova.compute.manager [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] instance snapshotting [ 911.532779] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2785ecc2-023d-4959-98dc-e619c8843d98 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.552146] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21d01d1-0fd6-459a-b11f-6fa7299fccc7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.601075] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.644420] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='dc8f8e1e6ad96acf30770e038ecd5965',container_format='bare',created_at=2024-10-12T10:58:31Z,direct_url=,disk_format='vmdk',id=250a0014-d97e-44f5-8fee-cc58ffff714d,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1590463486-shelved',owner='8bd0e94b74d74f9898049c9cad364b5b',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-10-12T10:58:46Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.644780] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.645014] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.645341] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.645573] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.645797] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.646097] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.646920] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.646920] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.646920] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.647095] env[61824]: DEBUG nova.virt.hardware [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.647925] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b702cc62-a471-497b-9659-edfbb1023c96 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.655905] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30133f05-c086-4e97-8a08-b61b1a1c1f34 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.673251] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:5b:14', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a66cb8c8-5826-4e02-ad96-b2bbea85d23b', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.681014] env[61824]: DEBUG oslo.service.loopingcall [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.681311] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.681533] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-86c4d093-37b6-48b8-83f1-eb819ef23f16 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.700878] env[61824]: DEBUG nova.scheduler.client.report [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 911.708387] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.708387] env[61824]: value = "task-1275790" [ 911.708387] env[61824]: _type = "Task" [ 911.708387] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.708695] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275788, 'name': ReconfigVM_Task, 'duration_secs': 1.284963} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.709095] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance '09432e72-c10a-4a7a-93af-af0d371fb871' progress to 33 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 911.722573] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275790, 'name': CreateVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.944062] env[61824]: DEBUG oslo_vmware.api [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275782, 'name': ReconfigVM_Task, 'duration_secs': 5.792514} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.944456] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.944745] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Reconfigured VM to detach interface {{(pid=61824) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 912.063705] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Creating Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 912.063705] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-254ed9e2-812a-4861-abee-d47b4a73602c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.071292] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 912.071292] env[61824]: value = "task-1275791" [ 912.071292] env[61824]: _type = "Task" [ 912.071292] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.080239] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275791, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.126709] env[61824]: DEBUG nova.compute.manager [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.126912] env[61824]: DEBUG nova.compute.manager [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing instance network info cache due to event network-changed-a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 912.127143] env[61824]: DEBUG oslo_concurrency.lockutils [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.127331] env[61824]: DEBUG oslo_concurrency.lockutils [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.127518] env[61824]: DEBUG nova.network.neutron [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Refreshing network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 912.205223] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.205744] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 912.208703] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.546s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.210162] env[61824]: INFO nova.compute.claims [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.216415] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.216647] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.216807] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.216990] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.217155] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.217306] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.217503] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.217667] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.217835] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.218046] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.218249] env[61824]: DEBUG nova.virt.hardware [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.223607] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Reconfiguring VM instance instance-00000050 to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 912.227084] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0494a1fe-1115-4de5-a4fa-19edce1a9871 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.245719] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275790, 'name': CreateVM_Task, 'duration_secs': 0.460961} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.246838] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.247189] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 912.247189] env[61824]: value = "task-1275792" [ 912.247189] env[61824]: _type = "Task" [ 912.247189] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.247922] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.248017] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.248371] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.249288] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b54884a-6733-4dcc-9731-51bc26ea2473 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.260166] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275792, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.261105] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 912.261105] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5240000c-5806-191b-eec4-941869992762" [ 912.261105] env[61824]: _type = "Task" [ 912.261105] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.269311] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5240000c-5806-191b-eec4-941869992762, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.582881] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275791, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.718032] env[61824]: DEBUG nova.compute.utils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 912.719090] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 912.719291] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 912.760455] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275792, 'name': ReconfigVM_Task, 'duration_secs': 0.157106} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.760736] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Reconfigured VM instance instance-00000050 to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 912.761707] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10cbec6-7fbe-49bd-af39-7877d6d8a81b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.782952] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.783422] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Processing image 250a0014-d97e-44f5-8fee-cc58ffff714d {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.783696] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d/250a0014-d97e-44f5-8fee-cc58ffff714d.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.783848] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d/250a0014-d97e-44f5-8fee-cc58ffff714d.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.784048] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.792177] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 09432e72-c10a-4a7a-93af-af0d371fb871/09432e72-c10a-4a7a-93af-af0d371fb871.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.792434] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67524eaf-5dfb-4e86-91f1-15a8ad25d490 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.794677] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10899257-4504-409a-a99d-779eaf58aec3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.815472] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 912.815472] env[61824]: value = "task-1275793" [ 912.815472] env[61824]: _type = "Task" [ 912.815472] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.816701] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.816878] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.820751] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79a88c0d-a175-417f-8296-9a6017c2eafd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.829020] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 912.829020] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52bfb183-e3b6-5e92-a363-1617f2110eaa" [ 912.829020] env[61824]: _type = "Task" [ 912.829020] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.829616] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275793, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.839570] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bfb183-e3b6-5e92-a363-1617f2110eaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.844224] env[61824]: DEBUG nova.policy [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9d0be6723dc4a448d478ad03f736739', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9788c8ba1ad14fb7a3abe2bd128c3623', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.083969] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275791, 'name': CreateSnapshot_Task, 'duration_secs': 0.963182} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.084309] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Created Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 913.084963] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb51dda7-f27f-4200-8f97-66001249825f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.128151] env[61824]: DEBUG nova.network.neutron [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updated VIF entry in instance network info cache for port a66cb8c8-5826-4e02-ad96-b2bbea85d23b. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 913.128593] env[61824]: DEBUG nova.network.neutron [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.223166] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 913.337961] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275793, 'name': ReconfigVM_Task, 'duration_secs': 0.447218} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.343272] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 09432e72-c10a-4a7a-93af-af0d371fb871/09432e72-c10a-4a7a-93af-af0d371fb871.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.343546] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance '09432e72-c10a-4a7a-93af-af0d371fb871' progress to 50 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 913.354924] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Preparing fetch location {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 913.354924] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Fetch image to [datastore1] OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404/OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404.vmdk {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 913.354924] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Downloading stream optimized image 250a0014-d97e-44f5-8fee-cc58ffff714d to [datastore1] OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404/OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404.vmdk on the data store datastore1 as vApp {{(pid=61824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 913.355111] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Downloading image file data 250a0014-d97e-44f5-8fee-cc58ffff714d to the ESX as VM named 'OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404' {{(pid=61824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 913.401205] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Successfully created port: 4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.446821] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 913.446821] env[61824]: value = "resgroup-9" [ 913.446821] env[61824]: _type = "ResourcePool" [ 913.446821] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 913.447759] env[61824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a094a348-aef2-4d02-a851-cefc23969304 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.471684] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease: (returnval){ [ 913.471684] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5289262b-33c2-a6d9-ae04-42b30bdd20bc" [ 913.471684] env[61824]: _type = "HttpNfcLease" [ 913.471684] env[61824]: } obtained for vApp import into resource pool (val){ [ 913.471684] env[61824]: value = "resgroup-9" [ 913.471684] env[61824]: _type = "ResourcePool" [ 913.471684] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 913.471967] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the lease: (returnval){ [ 913.471967] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5289262b-33c2-a6d9-ae04-42b30bdd20bc" [ 913.471967] env[61824]: _type = "HttpNfcLease" [ 913.471967] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 913.480741] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.480741] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5289262b-33c2-a6d9-ae04-42b30bdd20bc" [ 913.480741] env[61824]: _type = "HttpNfcLease" [ 913.480741] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 913.527679] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.527939] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquired lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.528135] env[61824]: DEBUG nova.network.neutron [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.546984] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca53fe11-67cc-4f32-a940-a368f8ff907b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.555252] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6321f4-d68f-43a8-b605-bff12e1aece9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.590342] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc98126e-ccea-4974-8134-9697be01a038 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.602733] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Creating linked-clone VM from snapshot {{(pid=61824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 913.604040] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c7d89f5d-a7fa-40ae-b692-cb156e26ba71 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.611722] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efd6942-9b43-44d7-aa51-db522d8de76f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.617453] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 913.617453] env[61824]: value = "task-1275795" [ 913.617453] env[61824]: _type = "Task" [ 913.617453] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.628998] env[61824]: DEBUG nova.compute.provider_tree [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 913.630724] env[61824]: DEBUG oslo_concurrency.lockutils [req-6e554227-225e-49ea-aab5-2c9440c7b90c req-a79f344d-d9c6-451c-a5ee-dd95e293196e service nova] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.635520] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275795, 'name': CloneVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.851865] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ab8262-39f1-46fc-a59e-6e45feb0d3a2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.873645] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac806d2-ceae-4a56-be6c-a2d158e0d97d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.892014] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance '09432e72-c10a-4a7a-93af-af0d371fb871' progress to 67 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 913.982999] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 913.982999] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5289262b-33c2-a6d9-ae04-42b30bdd20bc" [ 913.982999] env[61824]: _type = "HttpNfcLease" [ 913.982999] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 914.104033] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.104387] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.104567] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.104801] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.104982] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.107599] env[61824]: INFO nova.compute.manager [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Terminating instance [ 914.110214] env[61824]: DEBUG nova.compute.manager [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 914.111691] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 914.111691] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7b1560-a9c9-4052-b4d8-09940bbbaf37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.124795] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 914.125494] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94d0e7d3-011c-4216-a5c1-0c66d7bfa820 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.133106] env[61824]: DEBUG nova.scheduler.client.report [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.137571] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275795, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.145023] env[61824]: DEBUG oslo_vmware.api [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 914.145023] env[61824]: value = "task-1275797" [ 914.145023] env[61824]: _type = "Task" [ 914.145023] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.153639] env[61824]: DEBUG oslo_vmware.api [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275797, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.237664] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 914.263434] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.263732] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.263913] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.264134] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.264308] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.264466] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.264682] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.264870] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.265014] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.265194] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.265383] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.266318] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bcf21d-ba69-47ce-8440-6730e70a1379 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.274637] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33d4223-b1c7-40e7-b847-22d2cab1a849 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.481363] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 914.481363] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5289262b-33c2-a6d9-ae04-42b30bdd20bc" [ 914.481363] env[61824]: _type = "HttpNfcLease" [ 914.481363] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 914.481676] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 914.481676] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5289262b-33c2-a6d9-ae04-42b30bdd20bc" [ 914.481676] env[61824]: _type = "HttpNfcLease" [ 914.481676] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 914.482398] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d85394-d548-44d0-b4ef-9b1f3170de27 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.489761] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281051c-d2c0-a4cd-82cb-d4fecc71a58c/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 914.489944] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281051c-d2c0-a4cd-82cb-d4fecc71a58c/disk-0.vmdk. {{(pid=61824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 914.554811] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fe5d6e3b-37f3-4559-b5e5-adfe79a35f3d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.623420] env[61824]: DEBUG nova.network.neutron [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Port 542a267a-6d8a-4233-8c5a-64d5b20c7f72 binding to destination host cpu-1 is already ACTIVE {{(pid=61824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 914.631365] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275795, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.644987] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.645561] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 914.652019] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.938s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.652019] env[61824]: INFO nova.compute.claims [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 914.661405] env[61824]: DEBUG oslo_vmware.api [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275797, 'name': PowerOffVM_Task, 'duration_secs': 0.180412} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.661629] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.661798] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.662064] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa0641be-1562-40f4-9632-1804554442aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.672484] env[61824]: INFO nova.network.neutron [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Port a608c7e2-51a6-4791-b00b-570fb8a353b0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 914.673503] env[61824]: DEBUG nova.network.neutron [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [{"id": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "address": "fa:16:3e:d4:95:fc", "network": {"id": "34841295-0562-4caa-8275-c3695a63d719", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-47932744-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.225", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0924bfa41b5e42d68818557d2b0ace7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6edb8eae-1113-49d0-84f7-9fd9f82b26fb", "external-id": "nsx-vlan-transportzone-493", "segmentation_id": 493, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd35af69-df", "ovs_interfaceid": "bd35af69-df49-4c0a-a798-57bd9c8f6fe4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.743022] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.743022] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.743022] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleting the datastore file [datastore1] aa9c9eb2-b411-4eca-9f8d-3df377ffe351 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.743022] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81a61a4a-0b03-49b0-990e-0297b204281c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.748899] env[61824]: DEBUG oslo_vmware.api [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 914.748899] env[61824]: value = "task-1275799" [ 914.748899] env[61824]: _type = "Task" [ 914.748899] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.760783] env[61824]: DEBUG oslo_vmware.api [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275799, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.141479] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275795, 'name': CloneVM_Task, 'duration_secs': 1.225086} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.148123] env[61824]: INFO nova.virt.vmwareapi.vmops [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Created linked-clone VM from snapshot [ 915.154842] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f929f5e-e4c6-4372-932f-fcd44d222cb9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.160864] env[61824]: DEBUG nova.compute.utils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.165119] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 915.165309] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 915.176996] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Releasing lock "refresh_cache-aa9c9eb2-b411-4eca-9f8d-3df377ffe351" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.179626] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Uploading image 391e2405-4491-4f9d-8fee-52dfeb7f207c {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 915.193953] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Destroying the VM {{(pid=61824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 915.195325] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bcc76b11-77a1-4513-bcba-08b552edba73 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.200105] env[61824]: DEBUG nova.compute.manager [req-f7f8e951-9d80-4344-b54b-2706513b4277 req-5b7877f3-9bdc-4162-907e-7e62c5fffcf7 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Received event network-vif-plugged-4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.200105] env[61824]: DEBUG oslo_concurrency.lockutils [req-f7f8e951-9d80-4344-b54b-2706513b4277 req-5b7877f3-9bdc-4162-907e-7e62c5fffcf7 service nova] Acquiring lock "e5825315-bfea-4eec-80ca-cb68d9748d22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.200105] env[61824]: DEBUG oslo_concurrency.lockutils [req-f7f8e951-9d80-4344-b54b-2706513b4277 req-5b7877f3-9bdc-4162-907e-7e62c5fffcf7 service nova] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.200247] env[61824]: DEBUG oslo_concurrency.lockutils [req-f7f8e951-9d80-4344-b54b-2706513b4277 req-5b7877f3-9bdc-4162-907e-7e62c5fffcf7 service nova] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.200493] env[61824]: DEBUG nova.compute.manager [req-f7f8e951-9d80-4344-b54b-2706513b4277 req-5b7877f3-9bdc-4162-907e-7e62c5fffcf7 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] No waiting events found dispatching network-vif-plugged-4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 915.200629] env[61824]: WARNING nova.compute.manager [req-f7f8e951-9d80-4344-b54b-2706513b4277 req-5b7877f3-9bdc-4162-907e-7e62c5fffcf7 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Received unexpected event network-vif-plugged-4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f for instance with vm_state building and task_state spawning. [ 915.210669] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 915.210669] env[61824]: value = "task-1275800" [ 915.210669] env[61824]: _type = "Task" [ 915.210669] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.220873] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275800, 'name': Destroy_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.258459] env[61824]: DEBUG oslo_vmware.api [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275799, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148176} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.258738] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.258950] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.259145] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.259401] env[61824]: INFO nova.compute.manager [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Took 1.15 seconds to destroy the instance on the hypervisor. [ 915.259700] env[61824]: DEBUG oslo.service.loopingcall [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.259897] env[61824]: DEBUG nova.compute.manager [-] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.259995] env[61824]: DEBUG nova.network.neutron [-] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 915.275011] env[61824]: DEBUG nova.policy [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9d0be6723dc4a448d478ad03f736739', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9788c8ba1ad14fb7a3abe2bd128c3623', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.552987] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Successfully updated port: 4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.670743] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.671054] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.671267] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.672857] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 915.684274] env[61824]: DEBUG oslo_concurrency.lockutils [None req-bea61ff8-bdeb-487c-a94e-6390658afc9a tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "interface-aa9c9eb2-b411-4eca-9f8d-3df377ffe351-a608c7e2-51a6-4791-b00b-570fb8a353b0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.369s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.694851] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.695147] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.695302] env[61824]: INFO nova.compute.manager [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Shelving [ 915.726418] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275800, 'name': Destroy_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.905483] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Completed reading data from the image iterator. {{(pid=61824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 915.905715] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281051c-d2c0-a4cd-82cb-d4fecc71a58c/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 915.906823] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc597a0-92fc-47ad-9228-a612a548b805 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.913277] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281051c-d2c0-a4cd-82cb-d4fecc71a58c/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 915.913474] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281051c-d2c0-a4cd-82cb-d4fecc71a58c/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 915.913677] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ead85dc7-baab-4a09-8f34-1fa6a61ed87f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.943993] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5d5518-3eb1-4725-8ffc-34b86f59efed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.949468] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5de225-e7da-4050-8523-26dcbac65376 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.983893] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da63a45d-b00f-4598-850f-2b8428a78cd7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.992319] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa31d72-7b09-44ac-8005-365d7b83c376 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.008545] env[61824]: DEBUG nova.compute.provider_tree [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.042433] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Successfully created port: 9d6224ad-50c0-432d-a05d-d93e52155741 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.061694] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "refresh_cache-e5825315-bfea-4eec-80ca-cb68d9748d22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.061756] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "refresh_cache-e5825315-bfea-4eec-80ca-cb68d9748d22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.064367] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 916.204075] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.204348] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a0421d9-6ab3-45f6-b23d-3206c230d000 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.213078] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 916.213078] env[61824]: value = "task-1275801" [ 916.213078] env[61824]: _type = "Task" [ 916.213078] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.227686] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275801, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.231242] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275800, 'name': Destroy_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.363286] env[61824]: DEBUG oslo_vmware.rw_handles [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5281051c-d2c0-a4cd-82cb-d4fecc71a58c/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 916.363511] env[61824]: INFO nova.virt.vmwareapi.images [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Downloaded image file data 250a0014-d97e-44f5-8fee-cc58ffff714d [ 916.364371] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1d626b-2181-4bed-9e23-5eb3746ebf30 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.380373] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7668c38f-2d29-4853-86a5-f587b3383bcf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.403910] env[61824]: INFO nova.virt.vmwareapi.images [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] The imported VM was unregistered [ 916.407436] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Caching image {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 916.407801] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating directory with path [datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.408162] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-151427b9-c386-4ca7-a72e-261143d5e461 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.419421] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created directory with path [datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.419651] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404/OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404.vmdk to [datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d/250a0014-d97e-44f5-8fee-cc58ffff714d.vmdk. {{(pid=61824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 916.419927] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-514972bb-6533-41d7-8c18-442058173cac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.426693] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 916.426693] env[61824]: value = "task-1275803" [ 916.426693] env[61824]: _type = "Task" [ 916.426693] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.434469] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275803, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.512021] env[61824]: DEBUG nova.scheduler.client.report [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.669180] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 916.689244] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 916.722018] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 916.722294] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 916.722460] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 916.722650] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 916.722801] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 916.722950] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 916.723179] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 916.723343] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 916.723513] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 916.723679] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 916.723853] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 916.728453] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e0ea5b-95e0-4fed-87bf-8083a9929490 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.734619] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275800, 'name': Destroy_Task, 'duration_secs': 1.410516} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.735588] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Destroyed the VM [ 916.735842] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Deleting Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 916.736858] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.737017] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.740304] env[61824]: DEBUG nova.network.neutron [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 916.742657] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d7317422-c1f7-46c6-b3d3-e24e218af405 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.753587] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275801, 'name': PowerOffVM_Task, 'duration_secs': 0.206464} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.754896] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 916.756231] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05476bca-206c-477d-9f76-d73581fcf277 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.763160] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d979cce3-d89d-4c93-86d3-2dedbd3fe5c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.768233] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 916.768233] env[61824]: value = "task-1275804" [ 916.768233] env[61824]: _type = "Task" [ 916.768233] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.801181] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9c2a5c-ac4d-4b64-b93f-17f389911fa2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.808178] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275804, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.938335] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275803, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.017029] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.017586] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.020558] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.336s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.020947] env[61824]: DEBUG nova.objects.instance [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lazy-loading 'resources' on Instance uuid 9f031313-766b-4696-9bc6-75ced078bfde {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.098325] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Updating instance_info_cache with network_info: [{"id": "4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f", "address": "fa:16:3e:c4:06:f7", "network": {"id": "7dfb187f-99b3-4d94-8be7-91178fdd13a8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1557730459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9788c8ba1ad14fb7a3abe2bd128c3623", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5efce30e-48dd-493a-a354-f562a8adf7af", "external-id": "nsx-vlan-transportzone-283", "segmentation_id": 283, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b393fd2-ef", "ovs_interfaceid": "4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.249033] env[61824]: DEBUG nova.compute.manager [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Received event network-changed-4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.249383] env[61824]: DEBUG nova.compute.manager [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Refreshing instance network info cache due to event network-changed-4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.249383] env[61824]: DEBUG oslo_concurrency.lockutils [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] Acquiring lock "refresh_cache-e5825315-bfea-4eec-80ca-cb68d9748d22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.281480] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275804, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.316789] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Creating Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 917.317389] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f523ef5d-7d59-4c84-95c3-365c5091198b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.326120] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 917.326120] env[61824]: value = "task-1275805" [ 917.326120] env[61824]: _type = "Task" [ 917.326120] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.338909] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275805, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.440993] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275803, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.527226] env[61824]: DEBUG nova.compute.utils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.529127] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 917.529317] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 917.580948] env[61824]: DEBUG nova.network.neutron [-] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.593162] env[61824]: DEBUG nova.network.neutron [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance_info_cache with network_info: [{"id": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "address": "fa:16:3e:ad:9e:74", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap542a267a-6d", "ovs_interfaceid": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.602332] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "refresh_cache-e5825315-bfea-4eec-80ca-cb68d9748d22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.602332] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Instance network_info: |[{"id": "4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f", "address": "fa:16:3e:c4:06:f7", "network": {"id": "7dfb187f-99b3-4d94-8be7-91178fdd13a8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1557730459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9788c8ba1ad14fb7a3abe2bd128c3623", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5efce30e-48dd-493a-a354-f562a8adf7af", "external-id": "nsx-vlan-transportzone-283", "segmentation_id": 283, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b393fd2-ef", "ovs_interfaceid": "4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 917.603673] env[61824]: DEBUG oslo_concurrency.lockutils [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] Acquired lock "refresh_cache-e5825315-bfea-4eec-80ca-cb68d9748d22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.603673] env[61824]: DEBUG nova.network.neutron [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Refreshing network info cache for port 4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.604254] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:06:f7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5efce30e-48dd-493a-a354-f562a8adf7af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.614820] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Creating folder: Project (9788c8ba1ad14fb7a3abe2bd128c3623). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.617389] env[61824]: DEBUG nova.policy [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9d0be6723dc4a448d478ad03f736739', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9788c8ba1ad14fb7a3abe2bd128c3623', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 917.623965] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0030c254-2996-4882-acb5-a70b3859fb16 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.638519] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Created folder: Project (9788c8ba1ad14fb7a3abe2bd128c3623) in parent group-v274074. [ 917.639413] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Creating folder: Instances. Parent ref: group-v274192. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.639413] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f34116f5-69fa-4a53-9708-80dc738d8927 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.651783] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Created folder: Instances in parent group-v274192. [ 917.651783] env[61824]: DEBUG oslo.service.loopingcall [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.651994] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.652249] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-112a85a2-04e8-4076-bae5-69151648ddba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.677533] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.677533] env[61824]: value = "task-1275808" [ 917.677533] env[61824]: _type = "Task" [ 917.677533] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.686210] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275808, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.781145] env[61824]: DEBUG oslo_vmware.api [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275804, 'name': RemoveSnapshot_Task, 'duration_secs': 0.71019} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.781565] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Deleted Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 917.838711] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275805, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.876129] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b9c5a4-065d-4b46-b563-7fef548f3ecf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.884796] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a38b956-282a-4884-bdd5-e3f51064f01f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.921074] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1376fcf2-5424-4f0c-bb9b-5e8704aa088b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.929677] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd53f66-69e2-4f9e-9285-b35529881d1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.944516] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275803, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.954714] env[61824]: DEBUG nova.compute.provider_tree [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.032924] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.086823] env[61824]: INFO nova.compute.manager [-] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Took 2.83 seconds to deallocate network for instance. [ 918.094567] env[61824]: DEBUG oslo_concurrency.lockutils [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.189998] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275808, 'name': CreateVM_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.286947] env[61824]: WARNING nova.compute.manager [None req-7a279c9a-5fb9-4f79-88a4-84dafc8976dc tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Image not found during snapshot: nova.exception.ImageNotFound: Image 391e2405-4491-4f9d-8fee-52dfeb7f207c could not be found. [ 918.336677] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275805, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.444399] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275803, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.458822] env[61824]: DEBUG nova.scheduler.client.report [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.569292] env[61824]: DEBUG nova.network.neutron [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Updated VIF entry in instance network info cache for port 4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.569745] env[61824]: DEBUG nova.network.neutron [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Updating instance_info_cache with network_info: [{"id": "4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f", "address": "fa:16:3e:c4:06:f7", "network": {"id": "7dfb187f-99b3-4d94-8be7-91178fdd13a8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1557730459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9788c8ba1ad14fb7a3abe2bd128c3623", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5efce30e-48dd-493a-a354-f562a8adf7af", "external-id": "nsx-vlan-transportzone-283", "segmentation_id": 283, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b393fd2-ef", "ovs_interfaceid": "4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.593498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.630447] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8804ccac-f228-4d5d-a673-9a45a5357c0b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.655308] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2df4380-83ea-4408-a12f-69f6b6c6ffd3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.659509] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.659735] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.659889] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 918.666647] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance '09432e72-c10a-4a7a-93af-af0d371fb871' progress to 83 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 918.690851] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275808, 'name': CreateVM_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.849061] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275805, 'name': CreateSnapshot_Task, 'duration_secs': 1.184557} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.849134] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Created Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 918.850014] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44121f68-659f-4b83-864f-6247a1816ac9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.893958] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Successfully created port: dd67ede5-03e0-486b-a85e-1f51824d2dcd {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.946124] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275803, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.966307] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.967109] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.899s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.967898] env[61824]: DEBUG nova.objects.instance [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lazy-loading 'resources' on Instance uuid d6576f45-35f6-48dd-9160-80fc53d8b6b2 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 918.989954] env[61824]: INFO nova.scheduler.client.report [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Deleted allocations for instance 9f031313-766b-4696-9bc6-75ced078bfde [ 919.043378] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.075027] env[61824]: DEBUG oslo_concurrency.lockutils [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] Releasing lock "refresh_cache-e5825315-bfea-4eec-80ca-cb68d9748d22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.075027] env[61824]: DEBUG nova.compute.manager [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Received event network-vif-deleted-bd35af69-df49-4c0a-a798-57bd9c8f6fe4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.075027] env[61824]: INFO nova.compute.manager [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Neutron deleted interface bd35af69-df49-4c0a-a798-57bd9c8f6fe4; detaching it from the instance and deleting it from the info cache [ 919.075027] env[61824]: DEBUG nova.network.neutron [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.076731] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.077093] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.077370] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.077655] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.077898] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.078159] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.078528] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.078807] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.079089] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.079364] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.079663] env[61824]: DEBUG nova.virt.hardware [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.080915] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9b89d2a-3e71-48a0-ac63-4e157f35aa8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.086996] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Successfully updated port: 9d6224ad-50c0-432d-a05d-d93e52155741 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 919.088114] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb6a38b1-b26d-4118-a56c-77835a1b1ee4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.100019] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c47abe-ca2d-4704-a33a-830c2950fd4a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.103802] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82cf1b9-c931-4c98-be0d-35990d3165b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.134501] env[61824]: DEBUG nova.compute.manager [req-bbe8dff3-f214-4ec2-ab56-b44abc3452b9 req-b6e0edf4-9018-411f-88a0-5310351908a2 service nova] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Detach interface failed, port_id=bd35af69-df49-4c0a-a798-57bd9c8f6fe4, reason: Instance aa9c9eb2-b411-4eca-9f8d-3df377ffe351 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 919.177929] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.177929] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7be4f92c-9aaa-43fb-87e0-5768bf7a0ca5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.184956] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 919.184956] env[61824]: value = "task-1275809" [ 919.184956] env[61824]: _type = "Task" [ 919.184956] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.193040] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275808, 'name': CreateVM_Task, 'duration_secs': 1.373678} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.193668] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 919.194466] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.194782] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.195248] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 919.199377] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64d48db1-65a8-4db1-ab56-01ae30e9112e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.201503] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275809, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.207020] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 919.207020] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52703c7c-f17c-8014-7357-994bde6c9f32" [ 919.207020] env[61824]: _type = "Task" [ 919.207020] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.213704] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52703c7c-f17c-8014-7357-994bde6c9f32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.231747] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.231747] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquired lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.231747] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Forcefully refreshing network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 919.284075] env[61824]: DEBUG nova.compute.manager [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Received event network-vif-plugged-9d6224ad-50c0-432d-a05d-d93e52155741 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.284075] env[61824]: DEBUG oslo_concurrency.lockutils [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] Acquiring lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.284075] env[61824]: DEBUG oslo_concurrency.lockutils [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.284075] env[61824]: DEBUG oslo_concurrency.lockutils [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.284075] env[61824]: DEBUG nova.compute.manager [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] No waiting events found dispatching network-vif-plugged-9d6224ad-50c0-432d-a05d-d93e52155741 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 919.284075] env[61824]: WARNING nova.compute.manager [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Received unexpected event network-vif-plugged-9d6224ad-50c0-432d-a05d-d93e52155741 for instance with vm_state building and task_state spawning. [ 919.284075] env[61824]: DEBUG nova.compute.manager [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Received event network-changed-9d6224ad-50c0-432d-a05d-d93e52155741 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.284075] env[61824]: DEBUG nova.compute.manager [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Refreshing instance network info cache due to event network-changed-9d6224ad-50c0-432d-a05d-d93e52155741. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.284075] env[61824]: DEBUG oslo_concurrency.lockutils [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] Acquiring lock "refresh_cache-16a84291-3bb9-49a6-9c24-ca399fd22e4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.284075] env[61824]: DEBUG oslo_concurrency.lockutils [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] Acquired lock "refresh_cache-16a84291-3bb9-49a6-9c24-ca399fd22e4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.284075] env[61824]: DEBUG nova.network.neutron [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Refreshing network info cache for port 9d6224ad-50c0-432d-a05d-d93e52155741 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 919.375852] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Creating linked-clone VM from snapshot {{(pid=61824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 919.376628] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4a94d72e-2053-4800-b562-17e43ad5f030 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.391128] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 919.391128] env[61824]: value = "task-1275810" [ 919.391128] env[61824]: _type = "Task" [ 919.391128] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.398498] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275810, 'name': CloneVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.448192] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275803, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.564767} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.448192] env[61824]: INFO nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404/OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404.vmdk to [datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d/250a0014-d97e-44f5-8fee-cc58ffff714d.vmdk. [ 919.448192] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Cleaning up location [datastore1] OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 919.448192] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_96aa332e-6b3f-4606-a905-839fff8d2404 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.448192] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d5bb9d3b-8662-477a-9167-343525aad736 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.455061] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 919.455061] env[61824]: value = "task-1275811" [ 919.455061] env[61824]: _type = "Task" [ 919.455061] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.463350] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275811, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.502736] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d61fa95d-d354-4696-88ec-d9c3e423436b tempest-ServerMetadataNegativeTestJSON-669847662 tempest-ServerMetadataNegativeTestJSON-669847662-project-member] Lock "9f031313-766b-4696-9bc6-75ced078bfde" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.765s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.590953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "refresh_cache-16a84291-3bb9-49a6-9c24-ca399fd22e4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.699434] env[61824]: DEBUG oslo_vmware.api [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275809, 'name': PowerOnVM_Task, 'duration_secs': 0.412534} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.699759] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.700418] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-33229bd4-b932-41b2-bb57-f23a1e838d55 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance '09432e72-c10a-4a7a-93af-af0d371fb871' progress to 100 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 919.721522] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52703c7c-f17c-8014-7357-994bde6c9f32, 'name': SearchDatastore_Task, 'duration_secs': 0.018912} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.722583] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.722964] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.723081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.723201] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.723381] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.724134] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-defa1f1d-73b9-4804-81db-0efcdd4bcb46 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.726766] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f8cc654-cd3f-4fe0-8478-c7aeb74dd6cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.736606] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4931e63-0a89-4d5e-aede-6e7d86987958 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.741935] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.742148] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.767427] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f40042ca-34c1-497d-ab71-e77aa4384c6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.775037] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f4747d-a4d3-40e1-84a4-03e8217c3d9e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.781607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "beafe3e0-ff41-40f7-90cf-1d7977191180" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.781857] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.782072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "beafe3e0-ff41-40f7-90cf-1d7977191180-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.782256] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.782457] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.786370] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a5592f-47b4-495b-aa56-b6bb608f5f1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.792398] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 919.792398] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a92a6d-3fb6-7c26-7e21-ac5f7bf4d385" [ 919.792398] env[61824]: _type = "Task" [ 919.792398] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.792858] env[61824]: INFO nova.compute.manager [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Terminating instance [ 919.797885] env[61824]: DEBUG nova.compute.manager [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.798102] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.807735] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6963b8-ff83-42de-8fd9-7ef85b695871 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.812072] env[61824]: DEBUG nova.compute.provider_tree [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.817750] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a92a6d-3fb6-7c26-7e21-ac5f7bf4d385, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.819983] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.820781] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1cc6ab2-c905-48bb-b4ff-da8522ae653b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.829023] env[61824]: DEBUG oslo_vmware.api [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 919.829023] env[61824]: value = "task-1275812" [ 919.829023] env[61824]: _type = "Task" [ 919.829023] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.837011] env[61824]: DEBUG nova.network.neutron [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.841657] env[61824]: DEBUG oslo_vmware.api [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.906260] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275810, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.964238] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275811, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089143} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.964491] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.964666] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d/250a0014-d97e-44f5-8fee-cc58ffff714d.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.964927] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d/250a0014-d97e-44f5-8fee-cc58ffff714d.vmdk to [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.965307] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d0c226a5-d45d-4243-92a4-1275f89f7ee8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.974078] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 919.974078] env[61824]: value = "task-1275813" [ 919.974078] env[61824]: _type = "Task" [ 919.974078] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.983704] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.025800] env[61824]: DEBUG nova.network.neutron [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.310322] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a92a6d-3fb6-7c26-7e21-ac5f7bf4d385, 'name': SearchDatastore_Task, 'duration_secs': 0.060191} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.311937] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fde2e33-0010-42fa-8e2d-d6ec9d2b3bc3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.317256] env[61824]: DEBUG nova.scheduler.client.report [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 920.329145] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 920.329145] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5259f3bb-2b04-f37c-fc38-4eef88b43ac3" [ 920.329145] env[61824]: _type = "Task" [ 920.329145] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.357843] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5259f3bb-2b04-f37c-fc38-4eef88b43ac3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.358354] env[61824]: DEBUG oslo_vmware.api [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275812, 'name': PowerOffVM_Task, 'duration_secs': 0.245707} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.358891] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 920.359225] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 920.359631] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be2cec35-b7bb-478d-9168-d9cc715c1adc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.407027] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275810, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.457380] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 920.457380] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 920.457380] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleting the datastore file [datastore2] beafe3e0-ff41-40f7-90cf-1d7977191180 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.457966] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f06545d0-c60c-40ee-bd71-d9e1a775ca09 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.470202] env[61824]: DEBUG oslo_vmware.api [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for the task: (returnval){ [ 920.470202] env[61824]: value = "task-1275815" [ 920.470202] env[61824]: _type = "Task" [ 920.470202] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.487810] env[61824]: DEBUG oslo_vmware.api [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.495508] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275813, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.529543] env[61824]: DEBUG oslo_concurrency.lockutils [req-73c6957d-532f-42ed-b4b8-08456bc9c372 req-6cddf699-e5b3-4f90-b6d8-f3ff5089e453 service nova] Releasing lock "refresh_cache-16a84291-3bb9-49a6-9c24-ca399fd22e4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.530247] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "refresh_cache-16a84291-3bb9-49a6-9c24-ca399fd22e4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.530401] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 920.715071] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Updating instance_info_cache with network_info: [{"id": "64ce8844-4f88-487b-8fba-e00413a3eb05", "address": "fa:16:3e:05:3e:0f", "network": {"id": "3b0e8633-16fe-42c2-aefe-e98b0ab95ce7", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1750141297-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bad9de0b032c47429ba879d2d62c6af6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6076d24d-3c8e-4bbb-ba96-a08fb27a73cc", "external-id": "nsx-vlan-transportzone-267", "segmentation_id": 267, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap64ce8844-4f", "ovs_interfaceid": "64ce8844-4f88-487b-8fba-e00413a3eb05", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.828492] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.858s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.829677] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.964s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.830201] env[61824]: DEBUG nova.objects.instance [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'resources' on Instance uuid 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.850029] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5259f3bb-2b04-f37c-fc38-4eef88b43ac3, 'name': SearchDatastore_Task, 'duration_secs': 0.066631} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.851049] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.851049] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] e5825315-bfea-4eec-80ca-cb68d9748d22/e5825315-bfea-4eec-80ca-cb68d9748d22.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.851266] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-babe3fc2-369f-453f-a470-e19d5868194b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.856122] env[61824]: INFO nova.scheduler.client.report [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted allocations for instance d6576f45-35f6-48dd-9160-80fc53d8b6b2 [ 920.868663] env[61824]: DEBUG nova.compute.manager [req-60c61658-9512-4357-9b5b-dadd287a95d1 req-1a17d9fe-bc93-4a28-8f24-9f877f248fcf service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Received event network-vif-plugged-dd67ede5-03e0-486b-a85e-1f51824d2dcd {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.869363] env[61824]: DEBUG oslo_concurrency.lockutils [req-60c61658-9512-4357-9b5b-dadd287a95d1 req-1a17d9fe-bc93-4a28-8f24-9f877f248fcf service nova] Acquiring lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.869363] env[61824]: DEBUG oslo_concurrency.lockutils [req-60c61658-9512-4357-9b5b-dadd287a95d1 req-1a17d9fe-bc93-4a28-8f24-9f877f248fcf service nova] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.870586] env[61824]: DEBUG oslo_concurrency.lockutils [req-60c61658-9512-4357-9b5b-dadd287a95d1 req-1a17d9fe-bc93-4a28-8f24-9f877f248fcf service nova] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.870586] env[61824]: DEBUG nova.compute.manager [req-60c61658-9512-4357-9b5b-dadd287a95d1 req-1a17d9fe-bc93-4a28-8f24-9f877f248fcf service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] No waiting events found dispatching network-vif-plugged-dd67ede5-03e0-486b-a85e-1f51824d2dcd {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.870586] env[61824]: WARNING nova.compute.manager [req-60c61658-9512-4357-9b5b-dadd287a95d1 req-1a17d9fe-bc93-4a28-8f24-9f877f248fcf service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Received unexpected event network-vif-plugged-dd67ede5-03e0-486b-a85e-1f51824d2dcd for instance with vm_state building and task_state spawning. [ 920.872553] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 920.872553] env[61824]: value = "task-1275816" [ 920.872553] env[61824]: _type = "Task" [ 920.872553] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.885331] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275816, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.903038] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275810, 'name': CloneVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.982055] env[61824]: DEBUG oslo_vmware.api [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Task: {'id': task-1275815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307691} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.985761] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.985980] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 920.986207] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 920.986438] env[61824]: INFO nova.compute.manager [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Took 1.19 seconds to destroy the instance on the hypervisor. [ 920.986657] env[61824]: DEBUG oslo.service.loopingcall [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.987411] env[61824]: DEBUG nova.compute.manager [-] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 920.987529] env[61824]: DEBUG nova.network.neutron [-] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 920.996904] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275813, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.120605] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.220197] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Releasing lock "refresh_cache-7f6a4446-86e1-44c7-ab42-297b033dbace" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.220415] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Updated the network info_cache for instance {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 921.231037] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.231037] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.231882] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.231989] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.232371] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.232529] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.232666] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 921.232812] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.380049] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a7044fde-6648-47e6-882a-b217ece4bea1 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "d6576f45-35f6-48dd-9160-80fc53d8b6b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.346s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.385091] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Successfully updated port: dd67ede5-03e0-486b-a85e-1f51824d2dcd {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.394462] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275816, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.408742] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275810, 'name': CloneVM_Task, 'duration_secs': 1.638269} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.409068] env[61824]: INFO nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Created linked-clone VM from snapshot [ 921.409969] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d4aaf6-d04b-4b5b-97c4-59d598a8f492 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.420689] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Uploading image 0da722b8-2f9f-4a79-ada2-0ce4f4dd247e {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 921.455720] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 921.455720] env[61824]: value = "vm-274196" [ 921.455720] env[61824]: _type = "VirtualMachine" [ 921.455720] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 921.456055] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-468adfe2-5fdd-4353-aeda-621d75689816 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.469019] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lease: (returnval){ [ 921.469019] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6e1bd-543d-48da-c569-e29cc92708f2" [ 921.469019] env[61824]: _type = "HttpNfcLease" [ 921.469019] env[61824]: } obtained for exporting VM: (result){ [ 921.469019] env[61824]: value = "vm-274196" [ 921.469019] env[61824]: _type = "VirtualMachine" [ 921.469019] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 921.469282] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the lease: (returnval){ [ 921.469282] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6e1bd-543d-48da-c569-e29cc92708f2" [ 921.469282] env[61824]: _type = "HttpNfcLease" [ 921.469282] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 921.481872] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.481872] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6e1bd-543d-48da-c569-e29cc92708f2" [ 921.481872] env[61824]: _type = "HttpNfcLease" [ 921.481872] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 921.494981] env[61824]: DEBUG nova.compute.manager [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Received event network-changed-dd67ede5-03e0-486b-a85e-1f51824d2dcd {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.495235] env[61824]: DEBUG nova.compute.manager [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Refreshing instance network info cache due to event network-changed-dd67ede5-03e0-486b-a85e-1f51824d2dcd. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.495460] env[61824]: DEBUG oslo_concurrency.lockutils [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] Acquiring lock "refresh_cache-1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.495609] env[61824]: DEBUG oslo_concurrency.lockutils [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] Acquired lock "refresh_cache-1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.495772] env[61824]: DEBUG nova.network.neutron [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Refreshing network info cache for port dd67ede5-03e0-486b-a85e-1f51824d2dcd {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 921.503796] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275813, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.582520] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Updating instance_info_cache with network_info: [{"id": "9d6224ad-50c0-432d-a05d-d93e52155741", "address": "fa:16:3e:b8:fd:41", "network": {"id": "7dfb187f-99b3-4d94-8be7-91178fdd13a8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1557730459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9788c8ba1ad14fb7a3abe2bd128c3623", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5efce30e-48dd-493a-a354-f562a8adf7af", "external-id": "nsx-vlan-transportzone-283", "segmentation_id": 283, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d6224ad-50", "ovs_interfaceid": "9d6224ad-50c0-432d-a05d-d93e52155741", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.628090] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19613a5-6014-41c1-85c0-f5e1c27416a7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.637208] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f118b30c-ec7b-4ef6-9017-451168719b85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.673076] env[61824]: DEBUG nova.network.neutron [-] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.674848] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6033388-349e-4ef5-a6a3-1f141f1bfa3c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.685628] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb7e42e-6833-4e40-a5a1-22acb4b0323b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.702914] env[61824]: DEBUG nova.compute.provider_tree [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.737012] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.888970] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "refresh_cache-1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.894042] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275816, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.983655] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 921.983655] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6e1bd-543d-48da-c569-e29cc92708f2" [ 921.983655] env[61824]: _type = "HttpNfcLease" [ 921.983655] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 921.983655] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 921.983655] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c6e1bd-543d-48da-c569-e29cc92708f2" [ 921.983655] env[61824]: _type = "HttpNfcLease" [ 921.983655] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 921.983655] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eae9360-8864-4f98-93e5-51114379beb7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.996987] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275813, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.004065] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522d92ec-53d6-d5ad-f56e-53c930c223e9/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 922.004298] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522d92ec-53d6-d5ad-f56e-53c930c223e9/disk-0.vmdk for reading. {{(pid=61824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 922.068124] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "09432e72-c10a-4a7a-93af-af0d371fb871" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.068124] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.068124] env[61824]: DEBUG nova.compute.manager [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Going to confirm migration 2 {{(pid=61824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 922.074149] env[61824]: DEBUG nova.network.neutron [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 922.086908] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "refresh_cache-16a84291-3bb9-49a6-9c24-ca399fd22e4e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.087847] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Instance network_info: |[{"id": "9d6224ad-50c0-432d-a05d-d93e52155741", "address": "fa:16:3e:b8:fd:41", "network": {"id": "7dfb187f-99b3-4d94-8be7-91178fdd13a8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1557730459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9788c8ba1ad14fb7a3abe2bd128c3623", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5efce30e-48dd-493a-a354-f562a8adf7af", "external-id": "nsx-vlan-transportzone-283", "segmentation_id": 283, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d6224ad-50", "ovs_interfaceid": "9d6224ad-50c0-432d-a05d-d93e52155741", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.088122] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:fd:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5efce30e-48dd-493a-a354-f562a8adf7af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d6224ad-50c0-432d-a05d-d93e52155741', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.096099] env[61824]: DEBUG oslo.service.loopingcall [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.096774] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.097065] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6fe66d12-5be7-45c7-b2bd-5b1ecb2febb2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.118874] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-be636130-c32a-4b5a-bbd5-8f567b9f764a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.125244] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.125244] env[61824]: value = "task-1275818" [ 922.125244] env[61824]: _type = "Task" [ 922.125244] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.136124] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275818, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.160959] env[61824]: DEBUG nova.network.neutron [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.179052] env[61824]: INFO nova.compute.manager [-] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Took 1.19 seconds to deallocate network for instance. [ 922.205777] env[61824]: DEBUG nova.scheduler.client.report [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.331767] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "7f6a4446-86e1-44c7-ab42-297b033dbace" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.332199] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.332461] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "7f6a4446-86e1-44c7-ab42-297b033dbace-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.332689] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.332907] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.335185] env[61824]: INFO nova.compute.manager [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Terminating instance [ 922.337565] env[61824]: DEBUG nova.compute.manager [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.337769] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.338705] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5526bf71-e4ed-4814-a41f-95e7f22e778a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.347644] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.347878] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80f46ace-454b-4499-bbd6-00ac0cab075e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.356991] env[61824]: DEBUG oslo_vmware.api [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 922.356991] env[61824]: value = "task-1275819" [ 922.356991] env[61824]: _type = "Task" [ 922.356991] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.368138] env[61824]: DEBUG oslo_vmware.api [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275819, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.387316] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275816, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.500581] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275813, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.459396} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.500990] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/250a0014-d97e-44f5-8fee-cc58ffff714d/250a0014-d97e-44f5-8fee-cc58ffff714d.vmdk to [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.502060] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ec2a9a-5931-46ce-9b7b-ea399b63b226 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.529043] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.529675] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61bf1211-e735-446a-867b-2e5621131c90 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.554897] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 922.554897] env[61824]: value = "task-1275820" [ 922.554897] env[61824]: _type = "Task" [ 922.554897] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.567549] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275820, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.639030] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275818, 'name': CreateVM_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.640424] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.640793] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.641072] env[61824]: DEBUG nova.network.neutron [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.641280] env[61824]: DEBUG nova.objects.instance [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lazy-loading 'info_cache' on Instance uuid 09432e72-c10a-4a7a-93af-af0d371fb871 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.663769] env[61824]: DEBUG oslo_concurrency.lockutils [req-0202648b-9bb1-4e38-9c30-58ad7f13b94f req-6b895525-ec19-4f3c-b8a2-40add24f1df7 service nova] Releasing lock "refresh_cache-1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.664324] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "refresh_cache-1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.664560] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.686329] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.711772] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.882s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.718961] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.125s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.718961] env[61824]: DEBUG nova.objects.instance [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lazy-loading 'resources' on Instance uuid aa9c9eb2-b411-4eca-9f8d-3df377ffe351 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.748565] env[61824]: INFO nova.scheduler.client.report [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted allocations for instance 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5 [ 922.871294] env[61824]: DEBUG oslo_vmware.api [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275819, 'name': PowerOffVM_Task, 'duration_secs': 0.259174} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.871294] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.871294] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.871294] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-26629f62-ed55-4e19-a5f8-40449d2f8625 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.888087] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275816, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.956038} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.889668] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] e5825315-bfea-4eec-80ca-cb68d9748d22/e5825315-bfea-4eec-80ca-cb68d9748d22.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.889668] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.889668] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-93c5aec5-1d92-465d-8dea-4ab248504bb9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.900244] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 922.900244] env[61824]: value = "task-1275822" [ 922.900244] env[61824]: _type = "Task" [ 922.900244] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.912480] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275822, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.963605] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.964147] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.964571] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleting the datastore file [datastore1] 7f6a4446-86e1-44c7-ab42-297b033dbace {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.965033] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1ac5c39-b2fc-4da8-8ce6-ea987ffb6272 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.975675] env[61824]: DEBUG oslo_vmware.api [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 922.975675] env[61824]: value = "task-1275823" [ 922.975675] env[61824]: _type = "Task" [ 922.975675] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.985835] env[61824]: DEBUG oslo_vmware.api [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275823, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.066502] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275820, 'name': ReconfigVM_Task, 'duration_secs': 0.380491} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.067184] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2/7be79422-c624-4b4a-884f-29b13b9d81d2.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.068863] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'boot_index': 0, 'guest_format': None, 'encryption_secret_uuid': None, 'encryption_options': None, 'size': 0, 'encrypted': False, 'device_name': '/dev/sda', 'encryption_format': None, 'device_type': 'disk', 'image_id': '11b2670a-52f3-4a0e-8363-54ca7a7f5264'}], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'boot_index': None, 'guest_format': None, 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274181', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'name': 'volume-279b78fe-a845-4106-9b4f-158ad3979315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7be79422-c624-4b4a-884f-29b13b9d81d2', 'attached_at': '', 'detached_at': '', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'serial': '279b78fe-a845-4106-9b4f-158ad3979315'}, 'mount_device': '/dev/sdb', 'attachment_id': '8763ed34-759e-4802-8cb1-cace5a9e956f', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 923.069324] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 923.069749] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274181', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'name': 'volume-279b78fe-a845-4106-9b4f-158ad3979315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7be79422-c624-4b4a-884f-29b13b9d81d2', 'attached_at': '', 'detached_at': '', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'serial': '279b78fe-a845-4106-9b4f-158ad3979315'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 923.070642] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ed536e-7f50-454b-b6e3-e9fc916692a2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.090092] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c405e03-6296-4bff-8867-8a9a4f575f03 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.117043] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] volume-279b78fe-a845-4106-9b4f-158ad3979315/volume-279b78fe-a845-4106-9b4f-158ad3979315.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.117701] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88840c68-ad99-4469-b5eb-ff5b76f1ced6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.147029] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275818, 'name': CreateVM_Task, 'duration_secs': 0.962233} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.148714] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.149136] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 923.149136] env[61824]: value = "task-1275824" [ 923.149136] env[61824]: _type = "Task" [ 923.149136] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.149931] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.149931] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.150263] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.150710] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73569be1-2966-44f0-a203-5802581f67ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.166232] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275824, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.167307] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 923.167307] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52056033-0dbb-28bd-6fc8-78955ee5c3ce" [ 923.167307] env[61824]: _type = "Task" [ 923.167307] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.185331] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52056033-0dbb-28bd-6fc8-78955ee5c3ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.218884] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 923.256508] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4bde0b31-9dfa-4774-b53b-f0e31f2b3758 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "07973a77-0fe8-4bc9-b8c5-e693d7a63ef5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.913s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.411097] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275822, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096204} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.413829] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.415089] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65ebe1e-ccaa-4a91-b256-8d134ffd7699 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.439528] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] e5825315-bfea-4eec-80ca-cb68d9748d22/e5825315-bfea-4eec-80ca-cb68d9748d22.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.442686] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5e9ee58-a804-45aa-9064-4e6780a46be7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.459032] env[61824]: DEBUG nova.network.neutron [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Updating instance_info_cache with network_info: [{"id": "dd67ede5-03e0-486b-a85e-1f51824d2dcd", "address": "fa:16:3e:ba:47:86", "network": {"id": "7dfb187f-99b3-4d94-8be7-91178fdd13a8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1557730459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9788c8ba1ad14fb7a3abe2bd128c3623", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5efce30e-48dd-493a-a354-f562a8adf7af", "external-id": "nsx-vlan-transportzone-283", "segmentation_id": 283, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd67ede5-03", "ovs_interfaceid": "dd67ede5-03e0-486b-a85e-1f51824d2dcd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.470041] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 923.470041] env[61824]: value = "task-1275825" [ 923.470041] env[61824]: _type = "Task" [ 923.470041] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.483680] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275825, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.493521] env[61824]: DEBUG oslo_vmware.api [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275823, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.494872} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.493985] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.494287] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.494538] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.494838] env[61824]: INFO nova.compute.manager [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Took 1.16 seconds to destroy the instance on the hypervisor. [ 923.495208] env[61824]: DEBUG oslo.service.loopingcall [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.495655] env[61824]: DEBUG nova.compute.manager [-] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.495764] env[61824]: DEBUG nova.network.neutron [-] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.499916] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3fce7b-d2d4-4d51-904e-1261bca717a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.508795] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64cba88b-a51c-4c83-ba92-c966443ee1bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.543382] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2619b342-d21f-41fb-a931-4109cbfc7da6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.553590] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118c167a-7455-44b4-bf09-b4b5c8ce0a90 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.571041] env[61824]: DEBUG nova.compute.provider_tree [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.662756] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275824, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.678062] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52056033-0dbb-28bd-6fc8-78955ee5c3ce, 'name': SearchDatastore_Task, 'duration_secs': 0.02288} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.678454] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.678783] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.679094] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.679268] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.679589] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.679827] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d45e8031-0402-49c1-84ff-3fb0ac952413 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.693501] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.693800] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.695512] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3eafa13-e208-4592-9cb5-16a2016a188b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.707801] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 923.707801] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52ac7cdf-69d4-d6ab-1a82-da50b13f09f1" [ 923.707801] env[61824]: _type = "Task" [ 923.707801] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.725538] env[61824]: DEBUG nova.compute.manager [req-608c4bbb-bd27-4eb4-92af-5096a4a8e4fe req-69c9a6bb-e2dd-4806-b6b6-335d718cd223 service nova] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Received event network-vif-deleted-12e4609d-03ba-48c6-9526-a2c8e2c21b90 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.730742] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ac7cdf-69d4-d6ab-1a82-da50b13f09f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.963209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "refresh_cache-1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.963726] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Instance network_info: |[{"id": "dd67ede5-03e0-486b-a85e-1f51824d2dcd", "address": "fa:16:3e:ba:47:86", "network": {"id": "7dfb187f-99b3-4d94-8be7-91178fdd13a8", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1557730459-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9788c8ba1ad14fb7a3abe2bd128c3623", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5efce30e-48dd-493a-a354-f562a8adf7af", "external-id": "nsx-vlan-transportzone-283", "segmentation_id": 283, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd67ede5-03", "ovs_interfaceid": "dd67ede5-03e0-486b-a85e-1f51824d2dcd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.967018] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:47:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5efce30e-48dd-493a-a354-f562a8adf7af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd67ede5-03e0-486b-a85e-1f51824d2dcd', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.972797] env[61824]: DEBUG oslo.service.loopingcall [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.974075] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 923.977908] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-46db73db-6886-4e5d-8e0c-3d61612b6752 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.008713] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275825, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.011118] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.011118] env[61824]: value = "task-1275826" [ 924.011118] env[61824]: _type = "Task" [ 924.011118] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.020904] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275826, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.077068] env[61824]: DEBUG nova.scheduler.client.report [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.142523] env[61824]: DEBUG nova.network.neutron [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance_info_cache with network_info: [{"id": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "address": "fa:16:3e:ad:9e:74", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap542a267a-6d", "ovs_interfaceid": "542a267a-6d8a-4233-8c5a-64d5b20c7f72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.163701] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275824, 'name': ReconfigVM_Task, 'duration_secs': 0.65104} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.164096] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfigured VM instance instance-00000040 to attach disk [datastore1] volume-279b78fe-a845-4106-9b4f-158ad3979315/volume-279b78fe-a845-4106-9b4f-158ad3979315.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.170390] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e992aff-034e-4f1a-9f57-749ece82b7e7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.189658] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 924.189658] env[61824]: value = "task-1275827" [ 924.189658] env[61824]: _type = "Task" [ 924.189658] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.200442] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275827, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.219777] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ac7cdf-69d4-d6ab-1a82-da50b13f09f1, 'name': SearchDatastore_Task, 'duration_secs': 0.025888} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.220712] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfb2ec62-a876-4639-831b-d08f0f088ede {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.227516] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 924.227516] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d0b800-e87f-6c8c-13e5-aa01c4ece4b1" [ 924.227516] env[61824]: _type = "Task" [ 924.227516] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.237125] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d0b800-e87f-6c8c-13e5-aa01c4ece4b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.246847] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.247184] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.389583] env[61824]: DEBUG nova.network.neutron [-] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.484500] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275825, 'name': ReconfigVM_Task, 'duration_secs': 0.696521} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.484815] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Reconfigured VM instance instance-00000055 to attach disk [datastore1] e5825315-bfea-4eec-80ca-cb68d9748d22/e5825315-bfea-4eec-80ca-cb68d9748d22.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.485482] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1a890b1-17e2-445e-8c70-fe5415af673f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.493208] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 924.493208] env[61824]: value = "task-1275828" [ 924.493208] env[61824]: _type = "Task" [ 924.493208] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.501936] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275828, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.520885] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275826, 'name': CreateVM_Task, 'duration_secs': 0.423985} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.521328] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 924.521799] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.521966] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.522316] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.522581] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30770e09-a004-4c41-8940-dd30af3e5010 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.528386] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 924.528386] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5222f970-4f7a-7379-3cf4-55ab47a6be96" [ 924.528386] env[61824]: _type = "Task" [ 924.528386] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.538883] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5222f970-4f7a-7379-3cf4-55ab47a6be96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.584061] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.866s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.586827] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.850s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.586827] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.587094] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 924.587378] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.901s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.587602] env[61824]: DEBUG nova.objects.instance [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lazy-loading 'resources' on Instance uuid beafe3e0-ff41-40f7-90cf-1d7977191180 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.589298] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55680152-829e-4a47-b9cc-d4de57d5a7e4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.600510] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940b68b6-1031-4092-a714-a34211a4a9b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.617282] env[61824]: INFO nova.scheduler.client.report [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleted allocations for instance aa9c9eb2-b411-4eca-9f8d-3df377ffe351 [ 924.618787] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5466b2-9dce-41c3-9ec4-069db2071698 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.630430] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1b1aaf-24cb-4cc1-9bff-fdc078cadfe0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.664649] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-09432e72-c10a-4a7a-93af-af0d371fb871" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.664916] env[61824]: DEBUG nova.objects.instance [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lazy-loading 'migration_context' on Instance uuid 09432e72-c10a-4a7a-93af-af0d371fb871 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.666064] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179645MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 924.666207] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.700456] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275827, 'name': ReconfigVM_Task, 'duration_secs': 0.272797} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.700813] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274181', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'name': 'volume-279b78fe-a845-4106-9b4f-158ad3979315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7be79422-c624-4b4a-884f-29b13b9d81d2', 'attached_at': '', 'detached_at': '', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'serial': '279b78fe-a845-4106-9b4f-158ad3979315'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 924.701545] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d30eb4e-b39c-4898-bd22-402be1dd7591 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.710666] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 924.710666] env[61824]: value = "task-1275829" [ 924.710666] env[61824]: _type = "Task" [ 924.710666] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.721142] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275829, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.740932] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d0b800-e87f-6c8c-13e5-aa01c4ece4b1, 'name': SearchDatastore_Task, 'duration_secs': 0.013355} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.741304] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.741711] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 16a84291-3bb9-49a6-9c24-ca399fd22e4e/16a84291-3bb9-49a6-9c24-ca399fd22e4e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.742118] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b84838a-3427-46b2-a34d-79e31985ca62 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.750866] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 924.750866] env[61824]: value = "task-1275830" [ 924.750866] env[61824]: _type = "Task" [ 924.750866] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.751368] env[61824]: DEBUG nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.766081] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.893414] env[61824]: INFO nova.compute.manager [-] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Took 1.40 seconds to deallocate network for instance. [ 925.005629] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275828, 'name': Rename_Task, 'duration_secs': 0.175928} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.005629] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.005919] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17e5b5f4-2852-457f-a4a8-1ffa28239b97 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.014986] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 925.014986] env[61824]: value = "task-1275831" [ 925.014986] env[61824]: _type = "Task" [ 925.014986] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.025894] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275831, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.039168] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5222f970-4f7a-7379-3cf4-55ab47a6be96, 'name': SearchDatastore_Task, 'duration_secs': 0.013757} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.039616] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.039934] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.040330] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.040605] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.040888] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.041261] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88d7100c-aa77-4a27-b615-6ce9cd59ad4d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.058511] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.058896] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 925.060227] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a28d52e-1d8e-4897-9120-f2a7752c9c2f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.069415] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 925.069415] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5254738f-fcec-2410-9f68-cb2a78091e3e" [ 925.069415] env[61824]: _type = "Task" [ 925.069415] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.080050] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5254738f-fcec-2410-9f68-cb2a78091e3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.130700] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fd51a9dc-4e46-41be-852c-6f209890d911 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "aa9c9eb2-b411-4eca-9f8d-3df377ffe351" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.026s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.170024] env[61824]: DEBUG nova.objects.base [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Object Instance<09432e72-c10a-4a7a-93af-af0d371fb871> lazy-loaded attributes: info_cache,migration_context {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 925.170024] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1c90b9-dd2a-4264-9e94-de6451a8ef5f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.203605] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-849696a1-8973-4cef-a80c-8493f905776c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.211434] env[61824]: DEBUG oslo_vmware.api [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 925.211434] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5204287b-b38f-9cd6-c0e9-5a8ec969eddf" [ 925.211434] env[61824]: _type = "Task" [ 925.211434] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.233813] env[61824]: DEBUG oslo_vmware.api [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5204287b-b38f-9cd6-c0e9-5a8ec969eddf, 'name': SearchDatastore_Task, 'duration_secs': 0.015711} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.234527] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275829, 'name': Rename_Task, 'duration_secs': 0.169966} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.238589] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.238947] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.239475] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-344d7efa-0a28-4b81-afae-9e8c36ec6f7b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.249471] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 925.249471] env[61824]: value = "task-1275832" [ 925.249471] env[61824]: _type = "Task" [ 925.249471] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.269928] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.285502] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275830, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.290578] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.308892] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.310417] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.310417] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.310417] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.310417] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.312477] env[61824]: INFO nova.compute.manager [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Terminating instance [ 925.314854] env[61824]: DEBUG nova.compute.manager [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.315842] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 925.316131] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f6e3ce-9ec5-44c4-808f-d334a9f8479b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.325955] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.329786] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a62c5f7-d5ea-4390-9219-93be2ec8ea78 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.338369] env[61824]: DEBUG oslo_vmware.api [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 925.338369] env[61824]: value = "task-1275833" [ 925.338369] env[61824]: _type = "Task" [ 925.338369] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.350029] env[61824]: DEBUG oslo_vmware.api [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.358532] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d292d6-b932-4f46-9e82-173eb83e5b7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.367343] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d60d71b-9a73-4a46-ba0c-3fb5f0ff4635 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.400750] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.402530] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e218931-7cfe-4bc1-9fed-1a3bd013bf6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.412574] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59debcd7-234b-4e35-ac3d-b9895a8f0974 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.430633] env[61824]: DEBUG nova.compute.provider_tree [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.527375] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275831, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.581733] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5254738f-fcec-2410-9f68-cb2a78091e3e, 'name': SearchDatastore_Task, 'duration_secs': 0.059285} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.582622] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3622b7c-b53b-440c-adbe-0fd32f33e249 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.588781] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 925.588781] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]523ec60f-f63f-64ee-0616-d0f666d29f14" [ 925.588781] env[61824]: _type = "Task" [ 925.588781] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.597288] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523ec60f-f63f-64ee-0616-d0f666d29f14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.764994] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275832, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.767592] env[61824]: DEBUG nova.compute.manager [req-34ba3bcc-24d9-4fef-8f98-7ee582f80db0 req-2ae17eeb-c2b8-44d9-b547-89bacc0c33ae service nova] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Received event network-vif-deleted-64ce8844-4f88-487b-8fba-e00413a3eb05 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.771896] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275830, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561429} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.772107] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 16a84291-3bb9-49a6-9c24-ca399fd22e4e/16a84291-3bb9-49a6-9c24-ca399fd22e4e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.772321] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.772603] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3956682b-2a0a-4cec-8db6-fd8950ec7284 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.782030] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 925.782030] env[61824]: value = "task-1275834" [ 925.782030] env[61824]: _type = "Task" [ 925.782030] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.792056] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275834, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.851060] env[61824]: DEBUG oslo_vmware.api [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275833, 'name': PowerOffVM_Task, 'duration_secs': 0.257224} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.851406] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.851692] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.852043] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22529d57-7681-4d1e-9534-102fd885bac0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.923916] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.924197] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.924394] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleting the datastore file [datastore2] 5569eb0a-fd42-4e2d-b805-bb40e3264a65 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.924676] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b69009e9-e9fd-415a-8727-5aa12be38906 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.934661] env[61824]: DEBUG nova.scheduler.client.report [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.938443] env[61824]: DEBUG oslo_vmware.api [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for the task: (returnval){ [ 925.938443] env[61824]: value = "task-1275836" [ 925.938443] env[61824]: _type = "Task" [ 925.938443] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.948663] env[61824]: DEBUG oslo_vmware.api [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275836, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.030047] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275831, 'name': PowerOnVM_Task, 'duration_secs': 0.768363} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.030047] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.030047] env[61824]: INFO nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Took 11.79 seconds to spawn the instance on the hypervisor. [ 926.030047] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.030047] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b574835c-81b6-4eeb-b4cc-76ce0352252b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.103096] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523ec60f-f63f-64ee-0616-d0f666d29f14, 'name': SearchDatastore_Task, 'duration_secs': 0.013592} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.103489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.103885] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e/1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 926.104296] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80af3a2f-b316-4708-9aa8-8695e0b6b0b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.114036] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 926.114036] env[61824]: value = "task-1275837" [ 926.114036] env[61824]: _type = "Task" [ 926.114036] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.127828] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275837, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.262895] env[61824]: DEBUG oslo_vmware.api [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275832, 'name': PowerOnVM_Task, 'duration_secs': 0.52456} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.263898] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.293451] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275834, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071548} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.293592] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.294425] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68cccd1-9946-41cf-9f43-3ccb5f534ad7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.319289] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 16a84291-3bb9-49a6-9c24-ca399fd22e4e/16a84291-3bb9-49a6-9c24-ca399fd22e4e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.321094] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-618ba19f-2dab-45a7-a74a-91e34eee3fba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.342818] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 926.342818] env[61824]: value = "task-1275838" [ 926.342818] env[61824]: _type = "Task" [ 926.342818] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.352376] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275838, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.376209] env[61824]: DEBUG nova.compute.manager [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.376994] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e46196d-94cf-4d92-aafd-55874c6023f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.440118] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.853s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.443048] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.777s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.455434] env[61824]: DEBUG oslo_vmware.api [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Task: {'id': task-1275836, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326655} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.455716] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 926.455918] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 926.456147] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 926.456493] env[61824]: INFO nova.compute.manager [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Took 1.14 seconds to destroy the instance on the hypervisor. [ 926.456771] env[61824]: DEBUG oslo.service.loopingcall [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.456998] env[61824]: DEBUG nova.compute.manager [-] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 926.457111] env[61824]: DEBUG nova.network.neutron [-] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 926.467264] env[61824]: INFO nova.scheduler.client.report [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Deleted allocations for instance beafe3e0-ff41-40f7-90cf-1d7977191180 [ 926.550670] env[61824]: INFO nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Took 23.96 seconds to build instance. [ 926.631023] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275837, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.854774] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275838, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.898147] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fb5bdec0-06c6-458c-8ac2-65d38f1c6201 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 42.367s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.975153] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c63068c7-343a-4014-801e-19f99f89db16 tempest-ImagesTestJSON-1926769904 tempest-ImagesTestJSON-1926769904-project-member] Lock "beafe3e0-ff41-40f7-90cf-1d7977191180" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.193s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.053545] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.467s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.125779] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275837, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679693} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.126135] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e/1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 927.126417] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.126700] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7942e023-d8a0-47b6-8471-951bdae96a4c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.135597] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 927.135597] env[61824]: value = "task-1275839" [ 927.135597] env[61824]: _type = "Task" [ 927.135597] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.146138] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275839, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.354613] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275838, 'name': ReconfigVM_Task, 'duration_secs': 0.804328} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.354917] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 16a84291-3bb9-49a6-9c24-ca399fd22e4e/16a84291-3bb9-49a6-9c24-ca399fd22e4e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.355607] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1145b45-6cf8-424c-a485-36fd68aec9fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.366513] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 927.366513] env[61824]: value = "task-1275840" [ 927.366513] env[61824]: _type = "Task" [ 927.366513] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.376194] env[61824]: DEBUG nova.network.neutron [-] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.385136] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275840, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.456743] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Applying migration context for instance 09432e72-c10a-4a7a-93af-af0d371fb871 as it has an incoming, in-progress migration 208ec0f4-8708-472d-b4e9-bcd663ed20d5. Migration status is confirming {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 927.458132] env[61824]: INFO nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating resource usage from migration 208ec0f4-8708-472d-b4e9-bcd663ed20d5 [ 927.479852] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7e73f8b5-a138-4455-a392-9a2b2b860558 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.480048] env[61824]: WARNING nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7f6a4446-86e1-44c7-ab42-297b033dbace is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 927.480184] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 5569eb0a-fd42-4e2d-b805-bb40e3264a65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.480304] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.480420] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 21759f11-80e9-419d-9f65-ca4767d1593b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.480542] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7be79422-c624-4b4a-884f-29b13b9d81d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.480686] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.480803] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Migration 208ec0f4-8708-472d-b4e9-bcd663ed20d5 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 927.480957] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 09432e72-c10a-4a7a-93af-af0d371fb871 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.481104] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance e5825315-bfea-4eec-80ca-cb68d9748d22 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.481221] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 16a84291-3bb9-49a6-9c24-ca399fd22e4e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.481334] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.645863] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275839, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076579} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.646217] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.646909] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492f33df-68e3-43eb-b4f9-dd60ce3ead3c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.670399] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e/1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.670836] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-62880c0f-678e-44cf-ae37-e33bc14db65f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.703274] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 927.703274] env[61824]: value = "task-1275841" [ 927.703274] env[61824]: _type = "Task" [ 927.703274] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.714879] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275841, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.793507] env[61824]: DEBUG nova.compute.manager [req-7143c0e8-3d9a-476c-ac0c-02e0d86c4669 req-786f0ae5-6615-48f2-941f-42be92295484 service nova] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Received event network-vif-deleted-9baa0723-69a3-42a4-86c5-38ca3d336970 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.880031] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275840, 'name': Rename_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.886723] env[61824]: INFO nova.compute.manager [-] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Took 1.43 seconds to deallocate network for instance. [ 927.984564] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance c5da84f7-4ee7-4537-a3f5-b95ee513c4dd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 927.985050] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 927.985346] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 928.174229] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8533b019-1449-48f3-8f02-583ca882d999 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.182962] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6a83cb-2d75-4d23-92c9-908de9bbe1f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.220112] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ff2053-21d7-4ca8-a087-7454ac86f65e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.228590] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275841, 'name': ReconfigVM_Task, 'duration_secs': 0.418318} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.230855] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e/1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 928.231583] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c2882f8-6e60-4497-adb4-97e30be08674 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.234377] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84493d0c-51a5-4d6c-9818-483e5719aa92 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.250880] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.257718] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 928.257718] env[61824]: value = "task-1275842" [ 928.257718] env[61824]: _type = "Task" [ 928.257718] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.268547] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275842, 'name': Rename_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.381475] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275840, 'name': Rename_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.394577] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.765424] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.785352] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275842, 'name': Rename_Task, 'duration_secs': 0.171373} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.785352] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.785352] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5713a89c-633a-4f01-95c8-ede26b5119f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.793139] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 928.793139] env[61824]: value = "task-1275843" [ 928.793139] env[61824]: _type = "Task" [ 928.793139] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.803744] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275843, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.882319] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275840, 'name': Rename_Task, 'duration_secs': 1.335052} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.882319] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 928.882319] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ba55acd-48ad-412a-b475-110ce8832c86 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.893210] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 928.893210] env[61824]: value = "task-1275844" [ 928.893210] env[61824]: _type = "Task" [ 928.893210] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.910369] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275844, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.278337] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 929.278635] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.836s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.278932] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.041s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.309196] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275843, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.404458] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275844, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.810699] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275843, 'name': PowerOnVM_Task, 'duration_secs': 0.867828} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.811017] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.811293] env[61824]: INFO nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Took 10.77 seconds to spawn the instance on the hypervisor. [ 929.811559] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.812371] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e22af6c-5382-4cd7-a436-200984748e45 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.907702] env[61824]: DEBUG oslo_vmware.api [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275844, 'name': PowerOnVM_Task, 'duration_secs': 0.869674} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.908028] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 929.908240] env[61824]: INFO nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Took 13.22 seconds to spawn the instance on the hypervisor. [ 929.908501] env[61824]: DEBUG nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.909360] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a5319f-fc43-45b3-9059-588f60b1e9a2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.982609] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d944d9-5881-4362-bc7c-f86a7dcef505 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.991375] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec86135c-83b6-4e6f-9aae-38b7b067cc9e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.023829] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656e508b-85b5-4454-9640-90b0d662e446 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.032561] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628bb178-7edf-4d61-a788-ad982e32ef5f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.049351] env[61824]: DEBUG nova.compute.provider_tree [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.332254] env[61824]: INFO nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Took 27.64 seconds to build instance. [ 930.432406] env[61824]: INFO nova.compute.manager [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Took 27.79 seconds to build instance. [ 930.552230] env[61824]: DEBUG nova.scheduler.client.report [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.833950] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.152s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.935099] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cb74a7f2-e9f2-4315-8dc7-19a027bc60bd tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.304s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.988072] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522d92ec-53d6-d5ad-f56e-53c930c223e9/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 930.989138] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5882ff46-5fe5-4220-a5cd-5bccae18052b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.996559] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522d92ec-53d6-d5ad-f56e-53c930c223e9/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 930.996733] env[61824]: ERROR oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522d92ec-53d6-d5ad-f56e-53c930c223e9/disk-0.vmdk due to incomplete transfer. [ 930.996982] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-db4f8188-9aab-4bc1-bff4-7f3493b34508 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.005482] env[61824]: DEBUG oslo_vmware.rw_handles [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522d92ec-53d6-d5ad-f56e-53c930c223e9/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 931.005703] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Uploaded image 0da722b8-2f9f-4a79-ada2-0ce4f4dd247e to the Glance image server {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 931.008455] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Destroying the VM {{(pid=61824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 931.008871] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-922249fc-9d2b-4be3-9a4c-24b18846d93d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.016423] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 931.016423] env[61824]: value = "task-1275845" [ 931.016423] env[61824]: _type = "Task" [ 931.016423] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.026691] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275845, 'name': Destroy_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.528373] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275845, 'name': Destroy_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.562856] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.284s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.565587] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.275s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.567502] env[61824]: INFO nova.compute.claims [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.008159] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "e5825315-bfea-4eec-80ca-cb68d9748d22" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.008619] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.008763] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "e5825315-bfea-4eec-80ca-cb68d9748d22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.008980] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.009182] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.011751] env[61824]: INFO nova.compute.manager [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Terminating instance [ 932.013825] env[61824]: DEBUG nova.compute.manager [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.014038] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.014877] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a608bae-4372-42fb-b04b-a03c5afef4d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.026936] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.030145] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dd8e44b-cd4a-4902-98f6-48031aebbe1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.031739] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275845, 'name': Destroy_Task, 'duration_secs': 0.997057} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.032008] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Destroyed the VM [ 932.032260] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Deleting Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 932.032839] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c8554e52-d3a1-42f0-b107-810b21563d00 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.038882] env[61824]: DEBUG oslo_vmware.api [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 932.038882] env[61824]: value = "task-1275846" [ 932.038882] env[61824]: _type = "Task" [ 932.038882] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.042538] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 932.042538] env[61824]: value = "task-1275847" [ 932.042538] env[61824]: _type = "Task" [ 932.042538] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.052117] env[61824]: DEBUG oslo_vmware.api [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275846, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.058602] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275847, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.143548] env[61824]: INFO nova.scheduler.client.report [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted allocation for migration 208ec0f4-8708-472d-b4e9-bcd663ed20d5 [ 932.554630] env[61824]: DEBUG oslo_vmware.api [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275846, 'name': PowerOffVM_Task, 'duration_secs': 0.268099} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.559094] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 932.559350] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 932.559739] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275847, 'name': RemoveSnapshot_Task, 'duration_secs': 0.388661} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.559988] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f08f705f-9254-47aa-9d01-13866145af53 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.565025] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Deleted Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 932.565025] env[61824]: DEBUG nova.compute.manager [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.565025] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e5291e-0fd9-4c3c-a075-34544651da74 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.643087] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 932.644802] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 932.645083] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleting the datastore file [datastore1] e5825315-bfea-4eec-80ca-cb68d9748d22 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.648010] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14389ce3-c170-4484-8462-a9318a6bb128 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.653250] env[61824]: DEBUG oslo_concurrency.lockutils [None req-883779e0-27e8-40d8-90ce-1d36053c2a60 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.587s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.659022] env[61824]: DEBUG oslo_vmware.api [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 932.659022] env[61824]: value = "task-1275849" [ 932.659022] env[61824]: _type = "Task" [ 932.659022] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.667730] env[61824]: DEBUG oslo_vmware.api [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275849, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.786588] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560823bc-5b59-4834-947e-6524cec655ee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.795300] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2961db2b-5a84-4a7b-b879-9a44c79e822d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.827562] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e88b243-b5ea-4b83-a1a6-a838465b198d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.835913] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228c7679-16f0-48b1-bd37-1acff92bb3c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.853625] env[61824]: DEBUG nova.compute.provider_tree [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.081676] env[61824]: INFO nova.compute.manager [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Shelve offloading [ 933.086031] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.086031] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80ea6343-a227-41e8-8d15-7e9d3bcd81b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.095195] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 933.095195] env[61824]: value = "task-1275850" [ 933.095195] env[61824]: _type = "Task" [ 933.095195] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.109285] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 933.109624] env[61824]: DEBUG nova.compute.manager [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 933.110515] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc79503f-8892-4c81-a7fc-bae27770f09c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.118734] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.118888] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.118931] env[61824]: DEBUG nova.network.neutron [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.171325] env[61824]: DEBUG oslo_vmware.api [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275849, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196822} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.171603] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.171796] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.172113] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.172215] env[61824]: INFO nova.compute.manager [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Took 1.16 seconds to destroy the instance on the hypervisor. [ 933.172460] env[61824]: DEBUG oslo.service.loopingcall [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.172654] env[61824]: DEBUG nova.compute.manager [-] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 933.172751] env[61824]: DEBUG nova.network.neutron [-] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 933.356480] env[61824]: DEBUG nova.scheduler.client.report [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.556187] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "09432e72-c10a-4a7a-93af-af0d371fb871" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.556564] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.556834] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.557049] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.557229] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.559530] env[61824]: INFO nova.compute.manager [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Terminating instance [ 933.561885] env[61824]: DEBUG nova.compute.manager [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 933.562103] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.562948] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30669364-19e1-43bc-b048-f61a8649f2e5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.571918] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.572161] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ce71ee0-1bef-433d-bcc2-3b88b7dd68de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.582189] env[61824]: DEBUG oslo_vmware.api [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 933.582189] env[61824]: value = "task-1275851" [ 933.582189] env[61824]: _type = "Task" [ 933.582189] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.591353] env[61824]: DEBUG oslo_vmware.api [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.862408] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.863049] env[61824]: DEBUG nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.866656] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.466s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.866888] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.873240] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.476s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.873240] env[61824]: DEBUG nova.objects.instance [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lazy-loading 'resources' on Instance uuid 5569eb0a-fd42-4e2d-b805-bb40e3264a65 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.911251] env[61824]: INFO nova.scheduler.client.report [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted allocations for instance 7f6a4446-86e1-44c7-ab42-297b033dbace [ 934.052361] env[61824]: DEBUG nova.compute.manager [req-72b1eaed-4fd7-4e81-9fef-4b5bd65f0123 req-6ba2f4fc-7979-45d5-ba56-b5983c0aef1d service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Received event network-vif-deleted-4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.052361] env[61824]: INFO nova.compute.manager [req-72b1eaed-4fd7-4e81-9fef-4b5bd65f0123 req-6ba2f4fc-7979-45d5-ba56-b5983c0aef1d service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Neutron deleted interface 4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f; detaching it from the instance and deleting it from the info cache [ 934.052361] env[61824]: DEBUG nova.network.neutron [req-72b1eaed-4fd7-4e81-9fef-4b5bd65f0123 req-6ba2f4fc-7979-45d5-ba56-b5983c0aef1d service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.093850] env[61824]: DEBUG oslo_vmware.api [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275851, 'name': PowerOffVM_Task, 'duration_secs': 0.249154} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.094191] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 934.094361] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 934.094618] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45927469-f7b1-4031-8029-fbe02cfa75c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.180889] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 934.181234] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 934.181430] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleting the datastore file [datastore2] 09432e72-c10a-4a7a-93af-af0d371fb871 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.181704] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20334fd5-742e-4cb6-838c-de577f6802ac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.190418] env[61824]: DEBUG oslo_vmware.api [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 934.190418] env[61824]: value = "task-1275853" [ 934.190418] env[61824]: _type = "Task" [ 934.190418] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.203420] env[61824]: DEBUG oslo_vmware.api [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275853, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.250716] env[61824]: DEBUG nova.network.neutron [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45dc4201-9f", "ovs_interfaceid": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.373213] env[61824]: DEBUG nova.compute.utils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.374607] env[61824]: DEBUG nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.374801] env[61824]: DEBUG nova.network.neutron [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 934.424332] env[61824]: DEBUG oslo_concurrency.lockutils [None req-727c3ee7-dfae-4a71-a20c-0bb4ff48542f tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7f6a4446-86e1-44c7-ab42-297b033dbace" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.092s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.444653] env[61824]: DEBUG nova.policy [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1eb545ac2cf4f0d9ccddb040b17372b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e756c6817698464a9da391d4dd364946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.459258] env[61824]: DEBUG nova.network.neutron [-] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.517484] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "957ae086-5bb4-44f6-b287-5fbcd9742d12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.518103] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.555064] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-899f4e48-7228-4a73-822f-becc28c5b8c7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.568535] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee73e22-6428-459e-a215-7348de844810 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.603021] env[61824]: DEBUG nova.compute.manager [req-72b1eaed-4fd7-4e81-9fef-4b5bd65f0123 req-6ba2f4fc-7979-45d5-ba56-b5983c0aef1d service nova] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Detach interface failed, port_id=4b393fd2-efdb-4ad7-8dfd-f949c7c2b24f, reason: Instance e5825315-bfea-4eec-80ca-cb68d9748d22 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 934.605249] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4a4d31-d2bb-45e4-af28-14d3e8202eda {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.614051] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a927f14-5176-4b8a-a3ed-845e73a30097 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.646533] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab6d0ed-f31d-4e8e-9629-ae07ae4b0dfd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.657546] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ecc5de-a277-404f-ac14-afb705c9bb42 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.672287] env[61824]: DEBUG nova.compute.provider_tree [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.702178] env[61824]: DEBUG oslo_vmware.api [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275853, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.415556} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.702525] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.702723] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.702908] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.703098] env[61824]: INFO nova.compute.manager [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Took 1.14 seconds to destroy the instance on the hypervisor. [ 934.703362] env[61824]: DEBUG oslo.service.loopingcall [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.703557] env[61824]: DEBUG nova.compute.manager [-] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 934.703652] env[61824]: DEBUG nova.network.neutron [-] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 934.753460] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.831744] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "7e73f8b5-a138-4455-a392-9a2b2b860558" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.831892] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.832128] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "7e73f8b5-a138-4455-a392-9a2b2b860558-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.832320] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.832497] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.834792] env[61824]: INFO nova.compute.manager [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Terminating instance [ 934.836557] env[61824]: DEBUG nova.compute.manager [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 934.836773] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.837895] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-831fc77c-6bf7-4da7-87fa-992519508b38 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.846138] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.846389] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f24d6510-fa32-4e98-ba0e-474f3568d770 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.856369] env[61824]: DEBUG oslo_vmware.api [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 934.856369] env[61824]: value = "task-1275854" [ 934.856369] env[61824]: _type = "Task" [ 934.856369] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.863645] env[61824]: DEBUG oslo_vmware.api [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275854, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.879459] env[61824]: DEBUG nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.886190] env[61824]: DEBUG nova.network.neutron [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Successfully created port: a50e03ba-154c-4623-9e07-df8b06e3233e {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.961707] env[61824]: INFO nova.compute.manager [-] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Took 1.79 seconds to deallocate network for instance. [ 935.020230] env[61824]: DEBUG nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.174906] env[61824]: DEBUG nova.scheduler.client.report [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.376705] env[61824]: DEBUG oslo_vmware.api [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275854, 'name': PowerOffVM_Task, 'duration_secs': 0.306595} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.381017] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.381017] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.381017] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f6223cca-379f-46c4-8d22-0730747efd28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.471928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.474319] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.474654] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.474876] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleting the datastore file [datastore2] 7e73f8b5-a138-4455-a392-9a2b2b860558 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.475738] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ca6613a-371c-4c90-bbad-0ea0240c6279 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.485994] env[61824]: DEBUG oslo_vmware.api [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for the task: (returnval){ [ 935.485994] env[61824]: value = "task-1275856" [ 935.485994] env[61824]: _type = "Task" [ 935.485994] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.497141] env[61824]: DEBUG oslo_vmware.api [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275856, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.544717] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.685929] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.692623] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.216s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.692941] env[61824]: DEBUG nova.objects.instance [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lazy-loading 'resources' on Instance uuid e5825315-bfea-4eec-80ca-cb68d9748d22 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.708399] env[61824]: INFO nova.scheduler.client.report [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Deleted allocations for instance 5569eb0a-fd42-4e2d-b805-bb40e3264a65 [ 935.728909] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 935.728909] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023a3515-5a33-495f-b1a4-bf6c672b14ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.735709] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.736312] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bbd3a2e-07c8-4b1a-ab4f-2ea91d5643f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.820330] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.820330] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.820330] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleting the datastore file [datastore2] 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.820330] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d67d5d74-f513-4d90-b300-06fa3ba3a192 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.828623] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 935.828623] env[61824]: value = "task-1275858" [ 935.828623] env[61824]: _type = "Task" [ 935.828623] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.839267] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.893950] env[61824]: DEBUG nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.934633] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.934890] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.935130] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.935360] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.935525] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.935833] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.935903] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.936124] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.936348] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.936554] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.936910] env[61824]: DEBUG nova.virt.hardware [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.938051] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-681cb5ff-fa01-4386-b12a-9cff24bcfca2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.948739] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-465d71cd-0493-4c3c-a4df-8dfcd4491179 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.964489] env[61824]: DEBUG nova.network.neutron [-] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.997054] env[61824]: DEBUG oslo_vmware.api [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Task: {'id': task-1275856, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.397069} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.998034] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.998034] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.998034] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.998034] env[61824]: INFO nova.compute.manager [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Took 1.16 seconds to destroy the instance on the hypervisor. [ 935.998217] env[61824]: DEBUG oslo.service.loopingcall [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.998306] env[61824]: DEBUG nova.compute.manager [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.998393] env[61824]: DEBUG nova.network.neutron [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 936.098383] env[61824]: DEBUG nova.compute.manager [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Received event network-vif-deleted-542a267a-6d8a-4233-8c5a-64d5b20c7f72 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.098779] env[61824]: DEBUG nova.compute.manager [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received event network-vif-unplugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.098839] env[61824]: DEBUG oslo_concurrency.lockutils [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.099200] env[61824]: DEBUG oslo_concurrency.lockutils [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.099445] env[61824]: DEBUG oslo_concurrency.lockutils [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.099643] env[61824]: DEBUG nova.compute.manager [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] No waiting events found dispatching network-vif-unplugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.099890] env[61824]: WARNING nova.compute.manager [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received unexpected event network-vif-unplugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 for instance with vm_state shelved and task_state shelving_offloading. [ 936.100020] env[61824]: DEBUG nova.compute.manager [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received event network-changed-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.100220] env[61824]: DEBUG nova.compute.manager [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Refreshing instance network info cache due to event network-changed-45dc4201-9fb0-4060-ba24-f376e69a6e45. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 936.100419] env[61824]: DEBUG oslo_concurrency.lockutils [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] Acquiring lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.100564] env[61824]: DEBUG oslo_concurrency.lockutils [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] Acquired lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.100744] env[61824]: DEBUG nova.network.neutron [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Refreshing network info cache for port 45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 936.221115] env[61824]: DEBUG oslo_concurrency.lockutils [None req-75275d1e-a014-4832-bdfe-99813dbcefb8 tempest-AttachInterfacesTestJSON-1989277498 tempest-AttachInterfacesTestJSON-1989277498-project-member] Lock "5569eb0a-fd42-4e2d-b805-bb40e3264a65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.912s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.339486] env[61824]: DEBUG oslo_vmware.api [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.291208} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.339816] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.340025] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 936.340224] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 936.359500] env[61824]: INFO nova.scheduler.client.report [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleted allocations for instance 21759f11-80e9-419d-9f65-ca4767d1593b [ 936.380032] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe64bb8-371d-4294-b9b6-436a54d2a67f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.387371] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-098c0bd4-0810-4401-b4b0-09d533257b4f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.422985] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebe94be-2861-4679-ac6c-0a3e4affa17a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.434048] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f344f64-8f4f-450b-b8d6-1c601818681a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.448399] env[61824]: DEBUG nova.compute.provider_tree [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.466589] env[61824]: INFO nova.compute.manager [-] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Took 1.76 seconds to deallocate network for instance. [ 936.515530] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "e500f768-3ec8-48c7-a947-784e8665af15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.515530] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "e500f768-3ec8-48c7-a947-784e8665af15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.818825] env[61824]: DEBUG nova.network.neutron [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.864068] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.928341] env[61824]: DEBUG nova.network.neutron [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updated VIF entry in instance network info cache for port 45dc4201-9fb0-4060-ba24-f376e69a6e45. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.928767] env[61824]: DEBUG nova.network.neutron [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": null, "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap45dc4201-9f", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.952001] env[61824]: DEBUG nova.scheduler.client.report [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.974318] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.018698] env[61824]: DEBUG nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 937.192440] env[61824]: DEBUG nova.network.neutron [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Successfully updated port: a50e03ba-154c-4623-9e07-df8b06e3233e {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.322612] env[61824]: INFO nova.compute.manager [-] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Took 1.32 seconds to deallocate network for instance. [ 937.435406] env[61824]: DEBUG oslo_concurrency.lockutils [req-66ea6d14-b698-43dc-accc-32e10a092b8c req-2d6a147c-09ff-443e-ae1d-fce1e18396ea service nova] Releasing lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.461664] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.464951] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.920s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.466779] env[61824]: INFO nova.compute.claims [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.485977] env[61824]: INFO nova.scheduler.client.report [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleted allocations for instance e5825315-bfea-4eec-80ca-cb68d9748d22 [ 937.548176] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.694999] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "refresh_cache-c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.695158] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "refresh_cache-c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.695329] env[61824]: DEBUG nova.network.neutron [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 937.829319] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.996481] env[61824]: DEBUG oslo_concurrency.lockutils [None req-941bf3a0-e547-40be-8699-81c5592ba00e tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "e5825315-bfea-4eec-80ca-cb68d9748d22" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.988s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.038870] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.206218] env[61824]: DEBUG nova.compute.manager [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Received event network-vif-deleted-51fba888-670e-4f85-97a5-0133e0624342 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.206514] env[61824]: DEBUG nova.compute.manager [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Received event network-vif-plugged-a50e03ba-154c-4623-9e07-df8b06e3233e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.206786] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] Acquiring lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.207172] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.207413] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.207634] env[61824]: DEBUG nova.compute.manager [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] No waiting events found dispatching network-vif-plugged-a50e03ba-154c-4623-9e07-df8b06e3233e {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.207851] env[61824]: WARNING nova.compute.manager [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Received unexpected event network-vif-plugged-a50e03ba-154c-4623-9e07-df8b06e3233e for instance with vm_state building and task_state spawning. [ 938.208072] env[61824]: DEBUG nova.compute.manager [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Received event network-changed-a50e03ba-154c-4623-9e07-df8b06e3233e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.208276] env[61824]: DEBUG nova.compute.manager [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Refreshing instance network info cache due to event network-changed-a50e03ba-154c-4623-9e07-df8b06e3233e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.208495] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] Acquiring lock "refresh_cache-c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.251530] env[61824]: DEBUG nova.network.neutron [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 938.438608] env[61824]: DEBUG nova.network.neutron [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Updating instance_info_cache with network_info: [{"id": "a50e03ba-154c-4623-9e07-df8b06e3233e", "address": "fa:16:3e:9d:07:7e", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa50e03ba-15", "ovs_interfaceid": "a50e03ba-154c-4623-9e07-df8b06e3233e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.636044] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19158b9-71f7-403b-84a6-6fb6ae4ee8c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.644527] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1a7423-085b-4bea-88d8-46e76dab3289 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.675465] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e5678f-7dcc-4e9f-967b-a8138ec53b5c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.684542] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1a14507-f1d3-4ba7-aeec-d51c59ea831c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.699282] env[61824]: DEBUG nova.compute.provider_tree [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.925845] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.926443] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.926718] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.927161] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.927352] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.932894] env[61824]: INFO nova.compute.manager [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Terminating instance [ 938.934270] env[61824]: DEBUG nova.compute.manager [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.934463] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 938.935342] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc744b4a-39c5-43f2-a53c-a495f470dd3f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.940977] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "refresh_cache-c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.941335] env[61824]: DEBUG nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Instance network_info: |[{"id": "a50e03ba-154c-4623-9e07-df8b06e3233e", "address": "fa:16:3e:9d:07:7e", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa50e03ba-15", "ovs_interfaceid": "a50e03ba-154c-4623-9e07-df8b06e3233e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.942054] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] Acquired lock "refresh_cache-c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.942054] env[61824]: DEBUG nova.network.neutron [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Refreshing network info cache for port a50e03ba-154c-4623-9e07-df8b06e3233e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 938.943181] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:07:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a50e03ba-154c-4623-9e07-df8b06e3233e', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.950669] env[61824]: DEBUG oslo.service.loopingcall [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.953783] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.954361] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 938.954612] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3291d388-4851-4030-82aa-4e65f2db4ce7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.969503] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80dfeb86-a1f1-43fa-b244-de23ca2a45df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.979256] env[61824]: DEBUG oslo_vmware.api [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 938.979256] env[61824]: value = "task-1275859" [ 938.979256] env[61824]: _type = "Task" [ 938.979256] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.980550] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.980550] env[61824]: value = "task-1275860" [ 938.980550] env[61824]: _type = "Task" [ 938.980550] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.992979] env[61824]: DEBUG oslo_vmware.api [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275859, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.996058] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275860, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.998842] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.999134] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.999356] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.999551] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.999756] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.002752] env[61824]: INFO nova.compute.manager [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Terminating instance [ 939.005329] env[61824]: DEBUG nova.compute.manager [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 939.005533] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 939.006402] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082b6cd3-dcc5-4007-846f-c56adff1c16d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.016644] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 939.017067] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45774ad6-1b0e-44d7-aba1-e9264fb80a86 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.025671] env[61824]: DEBUG oslo_vmware.api [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 939.025671] env[61824]: value = "task-1275861" [ 939.025671] env[61824]: _type = "Task" [ 939.025671] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.038139] env[61824]: DEBUG oslo_vmware.api [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275861, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.204838] env[61824]: DEBUG nova.scheduler.client.report [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.493406] env[61824]: DEBUG oslo_vmware.api [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275859, 'name': PowerOffVM_Task, 'duration_secs': 0.295556} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.496852] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.497208] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.497347] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275860, 'name': CreateVM_Task, 'duration_secs': 0.472801} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.499625] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a349cf0-2dea-4f86-8303-d45e967c2525 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.501356] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 939.501772] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.501979] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.502343] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.503354] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68c3da96-57f8-40c7-a76f-e70a67bbe6e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.509521] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 939.509521] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a8201d-5bc9-0614-025e-823e78af831b" [ 939.509521] env[61824]: _type = "Task" [ 939.509521] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.518459] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a8201d-5bc9-0614-025e-823e78af831b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.536966] env[61824]: DEBUG oslo_vmware.api [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275861, 'name': PowerOffVM_Task, 'duration_secs': 0.289496} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.539371] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 939.539612] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 939.539951] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-39fefa57-919e-4986-8216-bfb6625fffac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.589585] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.590357] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.590571] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleting the datastore file [datastore1] 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.590966] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5df77b23-675a-4553-8cdd-1d43447e88f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.600490] env[61824]: DEBUG oslo_vmware.api [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 939.600490] env[61824]: value = "task-1275864" [ 939.600490] env[61824]: _type = "Task" [ 939.600490] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.609345] env[61824]: DEBUG oslo_vmware.api [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275864, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.620331] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 939.620520] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 939.620619] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleting the datastore file [datastore1] 16a84291-3bb9-49a6-9c24-ca399fd22e4e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.620908] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c1ca559-a5d2-4d12-835f-e6d2ee2d0301 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.629984] env[61824]: DEBUG oslo_vmware.api [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for the task: (returnval){ [ 939.629984] env[61824]: value = "task-1275865" [ 939.629984] env[61824]: _type = "Task" [ 939.629984] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.638370] env[61824]: DEBUG oslo_vmware.api [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275865, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.689512] env[61824]: DEBUG nova.network.neutron [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Updated VIF entry in instance network info cache for port a50e03ba-154c-4623-9e07-df8b06e3233e. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 939.689945] env[61824]: DEBUG nova.network.neutron [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Updating instance_info_cache with network_info: [{"id": "a50e03ba-154c-4623-9e07-df8b06e3233e", "address": "fa:16:3e:9d:07:7e", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa50e03ba-15", "ovs_interfaceid": "a50e03ba-154c-4623-9e07-df8b06e3233e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.709325] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.244s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.709898] env[61824]: DEBUG nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 939.712453] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.849s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.712814] env[61824]: DEBUG nova.objects.instance [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lazy-loading 'resources' on Instance uuid 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.020386] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a8201d-5bc9-0614-025e-823e78af831b, 'name': SearchDatastore_Task, 'duration_secs': 0.034495} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.020866] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.021270] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 940.021632] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.021928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.022245] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.022656] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-193b21ae-5fe8-4598-a11c-170bbdfac209 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.037693] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.038141] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 940.040091] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fbe4949-26bc-4965-8e41-54fc431991db {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.045878] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 940.045878] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5204cf89-c7a2-cc0e-4676-a813ad14574f" [ 940.045878] env[61824]: _type = "Task" [ 940.045878] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.055423] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5204cf89-c7a2-cc0e-4676-a813ad14574f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.111552] env[61824]: DEBUG oslo_vmware.api [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275864, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230569} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.112101] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.112500] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.113289] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.114232] env[61824]: INFO nova.compute.manager [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 940.114232] env[61824]: DEBUG oslo.service.loopingcall [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.114629] env[61824]: DEBUG nova.compute.manager [-] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.115926] env[61824]: DEBUG nova.network.neutron [-] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.141074] env[61824]: DEBUG oslo_vmware.api [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Task: {'id': task-1275865, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230783} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.141623] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.141830] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 940.142074] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 940.142478] env[61824]: INFO nova.compute.manager [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 940.142823] env[61824]: DEBUG oslo.service.loopingcall [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.143101] env[61824]: DEBUG nova.compute.manager [-] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.143234] env[61824]: DEBUG nova.network.neutron [-] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.192619] env[61824]: DEBUG oslo_concurrency.lockutils [req-ed6e2ebb-30fc-4ac6-b0b7-dba04ebeb3a2 req-2aabbac2-878b-4f71-959b-86a43ed55ff7 service nova] Releasing lock "refresh_cache-c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.217205] env[61824]: DEBUG nova.compute.utils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.217205] env[61824]: DEBUG nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 940.217205] env[61824]: DEBUG nova.network.neutron [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 940.221403] env[61824]: DEBUG nova.objects.instance [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lazy-loading 'numa_topology' on Instance uuid 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.543124] env[61824]: DEBUG nova.policy [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6a2f1ffe8934519acdcaca77e2a55d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e51dd2f578344aa8afa5d6beccb54f1e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.561176] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5204cf89-c7a2-cc0e-4676-a813ad14574f, 'name': SearchDatastore_Task, 'duration_secs': 0.017735} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.563398] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24e69f64-216e-4158-a6ba-5441b4ec19a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.571036] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 940.571036] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]524cd6fd-8b01-6dc5-fe20-43d53c6dc48a" [ 940.571036] env[61824]: _type = "Task" [ 940.571036] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.581677] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524cd6fd-8b01-6dc5-fe20-43d53c6dc48a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.721638] env[61824]: DEBUG nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 940.727658] env[61824]: DEBUG nova.objects.base [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Object Instance<21759f11-80e9-419d-9f65-ca4767d1593b> lazy-loaded attributes: resources,numa_topology {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 940.890982] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8272890e-357c-4af4-ac01-a656351051eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.902749] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b57d74-c0ae-4c08-b4d6-6c9017e33248 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.939467] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de0c6cb-9daa-447f-a1bf-ae52f1dda3a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.948778] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bf4f1b-9660-4462-a42b-747ee49de5ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.968821] env[61824]: DEBUG nova.compute.provider_tree [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.073258] env[61824]: DEBUG nova.compute.manager [req-ad49f942-6136-41e9-8783-9da7880636ba req-648664e2-f872-41ce-94c1-1d4475720581 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Received event network-vif-deleted-dd67ede5-03e0-486b-a85e-1f51824d2dcd {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 941.073258] env[61824]: INFO nova.compute.manager [req-ad49f942-6136-41e9-8783-9da7880636ba req-648664e2-f872-41ce-94c1-1d4475720581 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Neutron deleted interface dd67ede5-03e0-486b-a85e-1f51824d2dcd; detaching it from the instance and deleting it from the info cache [ 941.073258] env[61824]: DEBUG nova.network.neutron [req-ad49f942-6136-41e9-8783-9da7880636ba req-648664e2-f872-41ce-94c1-1d4475720581 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.086318] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524cd6fd-8b01-6dc5-fe20-43d53c6dc48a, 'name': SearchDatastore_Task, 'duration_secs': 0.034623} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.087501] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.087930] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] c5da84f7-4ee7-4537-a3f5-b95ee513c4dd/c5da84f7-4ee7-4537-a3f5-b95ee513c4dd.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 941.088350] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5870f679-9b97-40a6-8fe7-825fd3b4f428 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.096937] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 941.096937] env[61824]: value = "task-1275866" [ 941.096937] env[61824]: _type = "Task" [ 941.096937] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.105624] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.255047] env[61824]: DEBUG nova.network.neutron [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Successfully created port: 92b95e81-3307-478e-91c4-bfe7dd132b7d {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.446772] env[61824]: DEBUG nova.network.neutron [-] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.471768] env[61824]: DEBUG nova.scheduler.client.report [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.481188] env[61824]: DEBUG nova.network.neutron [-] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.576034] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa8157ef-ad1e-458f-b9bf-39cc55da23da {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.589020] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa82edd9-20e6-47ae-aed6-a7b99e72e9d1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.614067] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470091} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.614241] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] c5da84f7-4ee7-4537-a3f5-b95ee513c4dd/c5da84f7-4ee7-4537-a3f5-b95ee513c4dd.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.614432] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.614689] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf3f8b99-558f-4eaa-9719-421e809d15cf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.623192] env[61824]: DEBUG nova.compute.manager [req-ad49f942-6136-41e9-8783-9da7880636ba req-648664e2-f872-41ce-94c1-1d4475720581 service nova] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Detach interface failed, port_id=dd67ede5-03e0-486b-a85e-1f51824d2dcd, reason: Instance 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 941.625491] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 941.625491] env[61824]: value = "task-1275867" [ 941.625491] env[61824]: _type = "Task" [ 941.625491] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.633731] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.741151] env[61824]: DEBUG nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 941.778115] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.778385] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.778635] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.778934] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.779169] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.779386] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.779682] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.779908] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.780152] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.780679] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.780795] env[61824]: DEBUG nova.virt.hardware [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.781941] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dc12d4-0ca3-479e-bd94-3436dfd25c58 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.791573] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68412b8d-98c8-420b-964a-3577a37ef026 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.950593] env[61824]: INFO nova.compute.manager [-] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Took 1.84 seconds to deallocate network for instance. [ 941.983718] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.271s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.986697] env[61824]: INFO nova.compute.manager [-] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Took 1.84 seconds to deallocate network for instance. [ 941.986967] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.013s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.987182] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.989252] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.441s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.990717] env[61824]: INFO nova.compute.claims [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 942.016232] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "44ec5c22-d080-45d1-aa21-080af4c317c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.016460] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.018895] env[61824]: INFO nova.scheduler.client.report [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted allocations for instance 09432e72-c10a-4a7a-93af-af0d371fb871 [ 942.135217] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066581} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.135531] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 942.136353] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17de3bae-88a3-415d-bc2e-039cb6d28f04 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.159666] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] c5da84f7-4ee7-4537-a3f5-b95ee513c4dd/c5da84f7-4ee7-4537-a3f5-b95ee513c4dd.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.160268] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-734bfa64-a307-4cfd-a276-5eb56582bea4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.182726] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 942.182726] env[61824]: value = "task-1275868" [ 942.182726] env[61824]: _type = "Task" [ 942.182726] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.193695] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275868, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.458553] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.502685] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32046d52-7e7c-4c41-a238-697646c95e9a tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.807s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.506561] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.506561] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.466s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.506561] env[61824]: INFO nova.compute.manager [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Unshelving [ 942.518465] env[61824]: DEBUG nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.531676] env[61824]: DEBUG oslo_concurrency.lockutils [None req-56363a99-59e3-428c-b9ae-a7a5d61adc46 tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "09432e72-c10a-4a7a-93af-af0d371fb871" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.975s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.642742] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.642742] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.693972] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275868, 'name': ReconfigVM_Task, 'duration_secs': 0.456718} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.694288] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Reconfigured VM instance instance-00000058 to attach disk [datastore2] c5da84f7-4ee7-4537-a3f5-b95ee513c4dd/c5da84f7-4ee7-4537-a3f5-b95ee513c4dd.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.694935] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba166757-f758-4290-9d0c-daf41988b4df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.702718] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 942.702718] env[61824]: value = "task-1275869" [ 942.702718] env[61824]: _type = "Task" [ 942.702718] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.711553] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275869, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.973461] env[61824]: DEBUG nova.network.neutron [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Successfully updated port: 92b95e81-3307-478e-91c4-bfe7dd132b7d {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.045252] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.145599] env[61824]: DEBUG nova.compute.utils [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.171367] env[61824]: DEBUG nova.compute.manager [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Received event network-vif-deleted-9d6224ad-50c0-432d-a05d-d93e52155741 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.171592] env[61824]: DEBUG nova.compute.manager [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Received event network-vif-plugged-92b95e81-3307-478e-91c4-bfe7dd132b7d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.171788] env[61824]: DEBUG oslo_concurrency.lockutils [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] Acquiring lock "957ae086-5bb4-44f6-b287-5fbcd9742d12-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.172728] env[61824]: DEBUG oslo_concurrency.lockutils [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.172940] env[61824]: DEBUG oslo_concurrency.lockutils [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.173217] env[61824]: DEBUG nova.compute.manager [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] No waiting events found dispatching network-vif-plugged-92b95e81-3307-478e-91c4-bfe7dd132b7d {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.173339] env[61824]: WARNING nova.compute.manager [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Received unexpected event network-vif-plugged-92b95e81-3307-478e-91c4-bfe7dd132b7d for instance with vm_state building and task_state spawning. [ 943.173487] env[61824]: DEBUG nova.compute.manager [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Received event network-changed-92b95e81-3307-478e-91c4-bfe7dd132b7d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.173649] env[61824]: DEBUG nova.compute.manager [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Refreshing instance network info cache due to event network-changed-92b95e81-3307-478e-91c4-bfe7dd132b7d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.173843] env[61824]: DEBUG oslo_concurrency.lockutils [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] Acquiring lock "refresh_cache-957ae086-5bb4-44f6-b287-5fbcd9742d12" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.173982] env[61824]: DEBUG oslo_concurrency.lockutils [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] Acquired lock "refresh_cache-957ae086-5bb4-44f6-b287-5fbcd9742d12" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.174156] env[61824]: DEBUG nova.network.neutron [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Refreshing network info cache for port 92b95e81-3307-478e-91c4-bfe7dd132b7d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 943.178162] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7121ddd8-5974-4757-8d6c-a2564615fce0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.186829] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a72a2fe-b063-4455-b933-65b8425b6244 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.222704] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f9a7ed-ed40-464e-a76a-7a2854ef5a72 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.230991] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275869, 'name': Rename_Task, 'duration_secs': 0.153079} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.233214] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.233565] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bef7ee4-5ce1-43d3-a775-048e673ef25f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.236013] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0864be74-0158-417d-b056-61f0adc162fb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.250414] env[61824]: DEBUG nova.compute.provider_tree [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.252615] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 943.252615] env[61824]: value = "task-1275870" [ 943.252615] env[61824]: _type = "Task" [ 943.252615] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.261270] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275870, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.475998] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "refresh_cache-957ae086-5bb4-44f6-b287-5fbcd9742d12" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.529358] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.648647] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.704123] env[61824]: DEBUG nova.network.neutron [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 943.754406] env[61824]: DEBUG nova.scheduler.client.report [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 943.766958] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275870, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.775820] env[61824]: DEBUG nova.network.neutron [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.266018] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.266018] env[61824]: DEBUG nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 944.267117] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.438s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.267544] env[61824]: DEBUG nova.objects.instance [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lazy-loading 'resources' on Instance uuid 7e73f8b5-a138-4455-a392-9a2b2b860558 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.280059] env[61824]: DEBUG oslo_concurrency.lockutils [req-cbf2d346-ecf9-44aa-a8ed-0e834c4a3eeb req-ce8fd69d-7b3e-4c0c-8676-b85b98ca38c8 service nova] Releasing lock "refresh_cache-957ae086-5bb4-44f6-b287-5fbcd9742d12" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.281289] env[61824]: DEBUG oslo_vmware.api [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275870, 'name': PowerOnVM_Task, 'duration_secs': 1.000205} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.282382] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "refresh_cache-957ae086-5bb4-44f6-b287-5fbcd9742d12" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.283194] env[61824]: DEBUG nova.network.neutron [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 944.284207] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.284463] env[61824]: INFO nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Took 8.39 seconds to spawn the instance on the hypervisor. [ 944.284718] env[61824]: DEBUG nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.285846] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d038803-dd03-4d5d-add2-97fb11bbca50 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.739866] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.740188] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.740438] env[61824]: INFO nova.compute.manager [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Attaching volume d0589f73-9c49-4f8e-b869-a0a016254459 to /dev/sdb [ 944.772157] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0471c8-5bd1-4960-b717-31692578f20a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.775742] env[61824]: DEBUG nova.compute.utils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 944.780435] env[61824]: DEBUG nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 944.780435] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 944.789732] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e546a071-da51-474d-be41-18f6f730e362 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.804946] env[61824]: INFO nova.compute.manager [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Took 19.54 seconds to build instance. [ 944.810235] env[61824]: DEBUG nova.virt.block_device [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Updating existing volume attachment record: 22ad25cc-b32e-4de7-9ab9-b3455c61dd2b {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 944.829026] env[61824]: DEBUG nova.network.neutron [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 944.855135] env[61824]: DEBUG nova.policy [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8835a9932a0c42d68e6640bcc3b4978c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '773a259f132a42479464e120df314bb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 944.934993] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9f161b-6bbc-4447-a106-a1089301d212 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.943394] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ac718b-3361-41d4-a4df-c2a8271752b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.975187] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4615c74-c8b3-4112-afab-82ae203b0604 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.983015] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf70a6a-9efe-44ac-bcb3-4767b4a4d967 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.997355] env[61824]: DEBUG nova.compute.provider_tree [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.119700] env[61824]: DEBUG nova.network.neutron [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Updating instance_info_cache with network_info: [{"id": "92b95e81-3307-478e-91c4-bfe7dd132b7d", "address": "fa:16:3e:7b:a9:c4", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b95e81-33", "ovs_interfaceid": "92b95e81-3307-478e-91c4-bfe7dd132b7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.278722] env[61824]: DEBUG nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 945.306845] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2af1e3ca-d111-437f-9cb9-68cb745fa37b tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.060s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.330553] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Successfully created port: 10d4ea44-d470-4139-96d0-4c880d036717 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.500978] env[61824]: DEBUG nova.scheduler.client.report [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 945.622714] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "refresh_cache-957ae086-5bb4-44f6-b287-5fbcd9742d12" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.623090] env[61824]: DEBUG nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Instance network_info: |[{"id": "92b95e81-3307-478e-91c4-bfe7dd132b7d", "address": "fa:16:3e:7b:a9:c4", "network": {"id": "c3b067b2-102b-4dc8-9952-8bb9c1e2bfc5", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1390438371-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e51dd2f578344aa8afa5d6beccb54f1e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d646f9d5-d2ad-4c22-bea5-85a965334de6", "external-id": "nsx-vlan-transportzone-606", "segmentation_id": 606, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92b95e81-33", "ovs_interfaceid": "92b95e81-3307-478e-91c4-bfe7dd132b7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 945.623867] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7b:a9:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd646f9d5-d2ad-4c22-bea5-85a965334de6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92b95e81-3307-478e-91c4-bfe7dd132b7d', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.632462] env[61824]: DEBUG oslo.service.loopingcall [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.632703] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.632978] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eb8f6ca6-bd01-4577-a655-a52f8b1061fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.662464] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.662464] env[61824]: value = "task-1275872" [ 945.662464] env[61824]: _type = "Task" [ 945.662464] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.671499] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275872, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.778115] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Successfully created port: 6fe29327-b1fc-40ca-9f06-6bc80a449526 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.007300] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.739s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.010167] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.552s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.010428] env[61824]: DEBUG nova.objects.instance [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lazy-loading 'resources' on Instance uuid 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.036031] env[61824]: INFO nova.scheduler.client.report [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Deleted allocations for instance 7e73f8b5-a138-4455-a392-9a2b2b860558 [ 946.173866] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275872, 'name': CreateVM_Task, 'duration_secs': 0.364584} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.174051] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 946.174755] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.174960] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.175359] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 946.175647] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfb0abb4-db8f-46f6-af3e-9337369fc4f0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.178118] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Successfully created port: 2296c391-a8f5-4322-ac9f-7059d9ccb541 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 946.183561] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 946.183561] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d29d56-8894-8b48-e2ba-8f62b761227f" [ 946.183561] env[61824]: _type = "Task" [ 946.183561] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.192546] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d29d56-8894-8b48-e2ba-8f62b761227f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.288180] env[61824]: DEBUG nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 946.322427] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 946.322817] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 946.322899] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.323029] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 946.323191] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.323343] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 946.323553] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 946.323718] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 946.323912] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 946.324414] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 946.324648] env[61824]: DEBUG nova.virt.hardware [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 946.325530] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3501925c-5f84-40d3-820f-6a217e742dbd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.335211] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1644a3-77f3-445b-862f-9b71699291d6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.548529] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3e0f4170-3713-40e6-8f8d-49f5bc44c934 tempest-ServersAdminTestJSON-1616556962 tempest-ServersAdminTestJSON-1616556962-project-member] Lock "7e73f8b5-a138-4455-a392-9a2b2b860558" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.716s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.590633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "3256627d-35c8-4b75-b4a8-390eda0a150e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.591261] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.667602] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6ea39f-1d01-4cdb-aa88-f42f160b7065 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.676455] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df01fc64-9022-41de-9e90-0db1d6131ef1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.713085] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d377db-67aa-48a7-ac6a-1ed7c5414c20 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.728516] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737d2d6b-0a32-421e-9948-9e94d3271394 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.733230] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d29d56-8894-8b48-e2ba-8f62b761227f, 'name': SearchDatastore_Task, 'duration_secs': 0.01177} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.733524] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.733759] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.733989] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.734159] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.734854] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.734951] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a08bfded-90ed-4ed4-9ab6-65e77602f55d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.744785] env[61824]: DEBUG nova.compute.provider_tree [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.746888] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.747135] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.747997] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cd09e6d-6977-42f6-be7f-adf1659fbd1e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.754482] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 946.754482] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52886572-0cb4-3553-d2bb-6243aa9e4fae" [ 946.754482] env[61824]: _type = "Task" [ 946.754482] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.763156] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52886572-0cb4-3553-d2bb-6243aa9e4fae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.093813] env[61824]: DEBUG nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 947.249897] env[61824]: DEBUG nova.scheduler.client.report [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.267739] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52886572-0cb4-3553-d2bb-6243aa9e4fae, 'name': SearchDatastore_Task, 'duration_secs': 0.010025} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.269245] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1d82d8c-a934-4f7b-9f26-91fb63dc0af1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.275673] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 947.275673] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52938217-e391-9a1b-66e4-7671dff0efe6" [ 947.275673] env[61824]: _type = "Task" [ 947.275673] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.284813] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52938217-e391-9a1b-66e4-7671dff0efe6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.621680] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.750398] env[61824]: DEBUG nova.compute.manager [req-fc77f565-50ee-4d77-9d3c-e422df96d88c req-f44d0c25-57c0-4eb3-98c9-7ce38f47f518 service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-vif-plugged-10d4ea44-d470-4139-96d0-4c880d036717 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.750624] env[61824]: DEBUG oslo_concurrency.lockutils [req-fc77f565-50ee-4d77-9d3c-e422df96d88c req-f44d0c25-57c0-4eb3-98c9-7ce38f47f518 service nova] Acquiring lock "e500f768-3ec8-48c7-a947-784e8665af15-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.750832] env[61824]: DEBUG oslo_concurrency.lockutils [req-fc77f565-50ee-4d77-9d3c-e422df96d88c req-f44d0c25-57c0-4eb3-98c9-7ce38f47f518 service nova] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.751016] env[61824]: DEBUG oslo_concurrency.lockutils [req-fc77f565-50ee-4d77-9d3c-e422df96d88c req-f44d0c25-57c0-4eb3-98c9-7ce38f47f518 service nova] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.751193] env[61824]: DEBUG nova.compute.manager [req-fc77f565-50ee-4d77-9d3c-e422df96d88c req-f44d0c25-57c0-4eb3-98c9-7ce38f47f518 service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] No waiting events found dispatching network-vif-plugged-10d4ea44-d470-4139-96d0-4c880d036717 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.751411] env[61824]: WARNING nova.compute.manager [req-fc77f565-50ee-4d77-9d3c-e422df96d88c req-f44d0c25-57c0-4eb3-98c9-7ce38f47f518 service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received unexpected event network-vif-plugged-10d4ea44-d470-4139-96d0-4c880d036717 for instance with vm_state building and task_state spawning. [ 947.756177] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.744s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.757261] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.253s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.757491] env[61824]: DEBUG nova.objects.instance [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lazy-loading 'resources' on Instance uuid 16a84291-3bb9-49a6-9c24-ca399fd22e4e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.777811] env[61824]: INFO nova.scheduler.client.report [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleted allocations for instance 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e [ 947.795640] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52938217-e391-9a1b-66e4-7671dff0efe6, 'name': SearchDatastore_Task, 'duration_secs': 0.009828} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.795929] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.796201] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 957ae086-5bb4-44f6-b287-5fbcd9742d12/957ae086-5bb4-44f6-b287-5fbcd9742d12.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.796473] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-973dcef1-d3bb-4b69-ac00-2e2b8e99051a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.808016] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 947.808016] env[61824]: value = "task-1275874" [ 947.808016] env[61824]: _type = "Task" [ 947.808016] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.816102] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275874, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.870463] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Successfully updated port: 10d4ea44-d470-4139-96d0-4c880d036717 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 948.290059] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1b39e3f1-4f9e-4ec2-a609-8585f5534acb tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.364s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.321913] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275874, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460904} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.322124] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 957ae086-5bb4-44f6-b287-5fbcd9742d12/957ae086-5bb4-44f6-b287-5fbcd9742d12.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 948.322346] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.322601] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9391a630-1768-4e38-a068-fca4301d2b85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.333212] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 948.333212] env[61824]: value = "task-1275875" [ 948.333212] env[61824]: _type = "Task" [ 948.333212] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.343465] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275875, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.411608] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140500b6-bbec-4a94-9594-2bdf30627d49 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.420968] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030ea86e-cd58-4b15-8f7c-3e0fc3c0720b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.457260] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb3c8887-6bac-4f1f-be3e-048afbd5f33a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.465022] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa93e4b6-41e8-4f08-9269-3956383e3b6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.478797] env[61824]: DEBUG nova.compute.provider_tree [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.844652] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275875, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.227288} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.844933] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.845967] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62a470f-45ef-4750-8178-8a52fd8dfffb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.871082] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 957ae086-5bb4-44f6-b287-5fbcd9742d12/957ae086-5bb4-44f6-b287-5fbcd9742d12.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.871396] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2a0bdc5-ba1e-4348-acf2-9e9436a2fcad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.894960] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 948.894960] env[61824]: value = "task-1275876" [ 948.894960] env[61824]: _type = "Task" [ 948.894960] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.905644] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275876, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.983432] env[61824]: DEBUG nova.scheduler.client.report [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.354260] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 949.354522] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274200', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'name': 'volume-d0589f73-9c49-4f8e-b869-a0a016254459', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'serial': 'd0589f73-9c49-4f8e-b869-a0a016254459'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 949.355436] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf80095-453d-44a7-8c9f-573d1a3982cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.372919] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373dbd1f-9efc-494d-973e-56272614fe3a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.398083] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] volume-d0589f73-9c49-4f8e-b869-a0a016254459/volume-d0589f73-9c49-4f8e-b869-a0a016254459.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 949.399019] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be4e7532-49fa-45a2-b12f-03965d6e8df1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.424789] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.426185] env[61824]: DEBUG oslo_vmware.api [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 949.426185] env[61824]: value = "task-1275877" [ 949.426185] env[61824]: _type = "Task" [ 949.426185] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.438257] env[61824]: DEBUG oslo_vmware.api [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275877, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.488706] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.491106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.446s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.492605] env[61824]: INFO nova.compute.claims [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 949.512061] env[61824]: INFO nova.scheduler.client.report [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Deleted allocations for instance 16a84291-3bb9-49a6-9c24-ca399fd22e4e [ 949.789385] env[61824]: DEBUG nova.compute.manager [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-changed-10d4ea44-d470-4139-96d0-4c880d036717 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 949.789662] env[61824]: DEBUG nova.compute.manager [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Refreshing instance network info cache due to event network-changed-10d4ea44-d470-4139-96d0-4c880d036717. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 949.789816] env[61824]: DEBUG oslo_concurrency.lockutils [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] Acquiring lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.790014] env[61824]: DEBUG oslo_concurrency.lockutils [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] Acquired lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.790150] env[61824]: DEBUG nova.network.neutron [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Refreshing network info cache for port 10d4ea44-d470-4139-96d0-4c880d036717 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 949.909569] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275876, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.938039] env[61824]: DEBUG oslo_vmware.api [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275877, 'name': ReconfigVM_Task, 'duration_secs': 0.463976} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.938345] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Reconfigured VM instance instance-00000053 to attach disk [datastore1] volume-d0589f73-9c49-4f8e-b869-a0a016254459/volume-d0589f73-9c49-4f8e-b869-a0a016254459.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.943459] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f357175-80f8-4d4f-aea8-7665f3e2b8b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.960555] env[61824]: DEBUG oslo_vmware.api [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 949.960555] env[61824]: value = "task-1275878" [ 949.960555] env[61824]: _type = "Task" [ 949.960555] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.969348] env[61824]: DEBUG oslo_vmware.api [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.019900] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2207de10-2fe9-4b0b-a94c-44c76107c3f4 tempest-ListServersNegativeTestJSON-57012181 tempest-ListServersNegativeTestJSON-57012181-project-member] Lock "16a84291-3bb9-49a6-9c24-ca399fd22e4e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.021s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.344195] env[61824]: DEBUG nova.network.neutron [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 950.354686] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Successfully updated port: 6fe29327-b1fc-40ca-9f06-6bc80a449526 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.411116] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275876, 'name': ReconfigVM_Task, 'duration_secs': 1.045663} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.411434] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 957ae086-5bb4-44f6-b287-5fbcd9742d12/957ae086-5bb4-44f6-b287-5fbcd9742d12.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 950.412077] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-063c2506-1cf2-4e96-a6fa-062803ce6cd1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.419782] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 950.419782] env[61824]: value = "task-1275879" [ 950.419782] env[61824]: _type = "Task" [ 950.419782] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.429889] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275879, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.449913] env[61824]: DEBUG nova.network.neutron [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.472899] env[61824]: DEBUG oslo_vmware.api [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275878, 'name': ReconfigVM_Task, 'duration_secs': 0.152476} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.473201] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274200', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'name': 'volume-d0589f73-9c49-4f8e-b869-a0a016254459', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'serial': 'd0589f73-9c49-4f8e-b869-a0a016254459'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 950.638006] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5f23ca-cab0-469b-8f4a-3bbd5fe7dc89 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.646608] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bfc2972-a018-4b50-bbe6-21eb696a7345 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.678015] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b511b3aa-e1ce-4b0f-b103-1cabb3df6091 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.686190] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c5d538-b462-4f6c-9c56-ccfedd5fe3ba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.700549] env[61824]: DEBUG nova.compute.provider_tree [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.934560] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275879, 'name': Rename_Task, 'duration_secs': 0.210242} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.934933] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.935807] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83527238-e380-419a-bcaa-11cf3a8852ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.945780] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 950.945780] env[61824]: value = "task-1275880" [ 950.945780] env[61824]: _type = "Task" [ 950.945780] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.959114] env[61824]: DEBUG oslo_concurrency.lockutils [req-4a04c03a-4ae5-4465-961d-74eb9db81f5a req-909028db-0a41-4d58-b894-7b122389d2ec service nova] Releasing lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.959114] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275880, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.203332] env[61824]: DEBUG nova.scheduler.client.report [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.462505] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275880, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.526154] env[61824]: DEBUG nova.objects.instance [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lazy-loading 'flavor' on Instance uuid 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.713463] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.222s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.714035] env[61824]: DEBUG nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 951.719257] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.190s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.719496] env[61824]: DEBUG nova.objects.instance [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lazy-loading 'pci_requests' on Instance uuid 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.848925] env[61824]: DEBUG nova.compute.manager [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-vif-plugged-6fe29327-b1fc-40ca-9f06-6bc80a449526 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.849178] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] Acquiring lock "e500f768-3ec8-48c7-a947-784e8665af15-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.849393] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.849564] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.849771] env[61824]: DEBUG nova.compute.manager [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] No waiting events found dispatching network-vif-plugged-6fe29327-b1fc-40ca-9f06-6bc80a449526 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.849981] env[61824]: WARNING nova.compute.manager [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received unexpected event network-vif-plugged-6fe29327-b1fc-40ca-9f06-6bc80a449526 for instance with vm_state building and task_state spawning. [ 951.850171] env[61824]: DEBUG nova.compute.manager [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-changed-6fe29327-b1fc-40ca-9f06-6bc80a449526 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.850333] env[61824]: DEBUG nova.compute.manager [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Refreshing instance network info cache due to event network-changed-6fe29327-b1fc-40ca-9f06-6bc80a449526. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 951.850629] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] Acquiring lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.850680] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] Acquired lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.850809] env[61824]: DEBUG nova.network.neutron [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Refreshing network info cache for port 6fe29327-b1fc-40ca-9f06-6bc80a449526 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.957608] env[61824]: DEBUG oslo_vmware.api [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275880, 'name': PowerOnVM_Task, 'duration_secs': 0.578295} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.957901] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 951.958129] env[61824]: INFO nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Took 10.22 seconds to spawn the instance on the hypervisor. [ 951.958321] env[61824]: DEBUG nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.959236] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37cdd3ef-3a52-4fc7-bd0a-70a7ec52ce2d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.032286] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1500dcd2-67a4-4fe6-9b48-ce86aed715f0 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.292s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.225972] env[61824]: DEBUG nova.compute.utils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 952.231544] env[61824]: DEBUG nova.objects.instance [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lazy-loading 'numa_topology' on Instance uuid 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.232701] env[61824]: DEBUG nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 952.232873] env[61824]: DEBUG nova.network.neutron [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 952.330618] env[61824]: DEBUG nova.policy [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8763f64818e24fad80ce232ac7d4463f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '629276280a234e7a8b810e6e74779fff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 952.332842] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.333100] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.443284] env[61824]: DEBUG nova.network.neutron [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 952.477546] env[61824]: INFO nova.compute.manager [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Took 16.95 seconds to build instance. [ 952.733792] env[61824]: DEBUG nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 952.741156] env[61824]: INFO nova.compute.claims [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.755239] env[61824]: DEBUG nova.network.neutron [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.837291] env[61824]: INFO nova.compute.manager [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Detaching volume d0589f73-9c49-4f8e-b869-a0a016254459 [ 952.895474] env[61824]: INFO nova.virt.block_device [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Attempting to driver detach volume d0589f73-9c49-4f8e-b869-a0a016254459 from mountpoint /dev/sdb [ 952.895730] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 952.896399] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274200', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'name': 'volume-d0589f73-9c49-4f8e-b869-a0a016254459', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'serial': 'd0589f73-9c49-4f8e-b869-a0a016254459'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 952.898198] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833304d7-d318-4202-985e-a715d124ceb1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.923339] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8c4012-71ae-4cc7-932d-3a8d67b81cd8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.931455] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043bb9ad-658f-4b05-b128-8ce3510e019b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.956671] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a11f7f0-c675-4329-9410-82f3e8690dac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.972578] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] The volume has not been displaced from its original location: [datastore1] volume-d0589f73-9c49-4f8e-b869-a0a016254459/volume-d0589f73-9c49-4f8e-b869-a0a016254459.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 952.978741] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Reconfiguring VM instance instance-00000053 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 952.979459] env[61824]: DEBUG oslo_concurrency.lockutils [None req-908123ba-86ab-4743-a7b2-d4ce411ae34f tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.462s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.979691] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9233e3a8-eba6-4ad9-9890-2c3bf19829e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.002065] env[61824]: DEBUG oslo_vmware.api [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 953.002065] env[61824]: value = "task-1275881" [ 953.002065] env[61824]: _type = "Task" [ 953.002065] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.009910] env[61824]: DEBUG oslo_vmware.api [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.258520] env[61824]: DEBUG oslo_concurrency.lockutils [req-eb3df660-4352-48e4-9951-401c83c17db0 req-b334a68d-7950-4ecb-8546-328db42af94e service nova] Releasing lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.267216] env[61824]: DEBUG nova.network.neutron [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Successfully created port: 7832e699-039f-4690-9e62-cdb49ebee107 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 953.494131] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Successfully updated port: 2296c391-a8f5-4322-ac9f-7059d9ccb541 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.515129] env[61824]: DEBUG oslo_vmware.api [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275881, 'name': ReconfigVM_Task, 'duration_secs': 0.243615} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.515410] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Reconfigured VM instance instance-00000053 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 953.520433] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb9a07d9-17c2-453f-8688-a1a0ca5a11de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.539395] env[61824]: DEBUG oslo_vmware.api [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 953.539395] env[61824]: value = "task-1275882" [ 953.539395] env[61824]: _type = "Task" [ 953.539395] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.545650] env[61824]: DEBUG oslo_vmware.api [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275882, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.747694] env[61824]: DEBUG nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 953.773282] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 953.773556] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 953.773816] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 953.777161] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 953.777161] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 953.777161] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 953.777161] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 953.777161] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 953.777161] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 953.777161] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 953.777398] env[61824]: DEBUG nova.virt.hardware [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 953.778482] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b15acb4-adac-4ae3-9722-a4c93b068b29 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.788956] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e67971-8289-4a0f-917a-da3ba0644f0f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.900859] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53db8e78-01b7-44d7-988d-b875d290f26b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.908207] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b61027f-1f37-4738-86de-c98c712c9e6f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.938977] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dadfd56-8db1-40c0-a8bd-c9f265ffec2b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.947163] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab653043-1220-4f2a-b9a2-af3c404f2626 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.960402] env[61824]: DEBUG nova.compute.provider_tree [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.997854] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.998051] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.998206] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 954.046947] env[61824]: DEBUG oslo_vmware.api [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275882, 'name': ReconfigVM_Task, 'duration_secs': 0.132425} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.047219] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274200', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'name': 'volume-d0589f73-9c49-4f8e-b869-a0a016254459', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd0589f73-9c49-4f8e-b869-a0a016254459', 'serial': 'd0589f73-9c49-4f8e-b869-a0a016254459'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 954.218131] env[61824]: DEBUG nova.compute.manager [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-vif-plugged-2296c391-a8f5-4322-ac9f-7059d9ccb541 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.218463] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] Acquiring lock "e500f768-3ec8-48c7-a947-784e8665af15-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.218803] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.219058] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.219294] env[61824]: DEBUG nova.compute.manager [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] No waiting events found dispatching network-vif-plugged-2296c391-a8f5-4322-ac9f-7059d9ccb541 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.219563] env[61824]: WARNING nova.compute.manager [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received unexpected event network-vif-plugged-2296c391-a8f5-4322-ac9f-7059d9ccb541 for instance with vm_state building and task_state spawning. [ 954.219770] env[61824]: DEBUG nova.compute.manager [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-changed-2296c391-a8f5-4322-ac9f-7059d9ccb541 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.220054] env[61824]: DEBUG nova.compute.manager [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Refreshing instance network info cache due to event network-changed-2296c391-a8f5-4322-ac9f-7059d9ccb541. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 954.220269] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] Acquiring lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.464341] env[61824]: DEBUG nova.scheduler.client.report [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.553914] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 954.594364] env[61824]: DEBUG nova.objects.instance [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lazy-loading 'flavor' on Instance uuid 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.832714] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "957ae086-5bb4-44f6-b287-5fbcd9742d12" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.832714] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.833157] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "957ae086-5bb4-44f6-b287-5fbcd9742d12-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.833157] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.833547] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.835814] env[61824]: INFO nova.compute.manager [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Terminating instance [ 954.838094] env[61824]: DEBUG nova.compute.manager [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 954.838297] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 954.839326] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36688fb1-6892-400d-9122-ac4db6b6fc4b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.847424] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.847622] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53b22df9-b156-4e15-8882-0ab2838c9a27 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.857112] env[61824]: DEBUG oslo_vmware.api [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 954.857112] env[61824]: value = "task-1275883" [ 954.857112] env[61824]: _type = "Task" [ 954.857112] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.971417] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.252s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.973384] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.352s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.975504] env[61824]: INFO nova.compute.claims [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.036447] env[61824]: INFO nova.network.neutron [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating port 45dc4201-9fb0-4060-ba24-f376e69a6e45 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 955.250482] env[61824]: DEBUG nova.network.neutron [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Updating instance_info_cache with network_info: [{"id": "10d4ea44-d470-4139-96d0-4c880d036717", "address": "fa:16:3e:65:a5:7f", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10d4ea44-d4", "ovs_interfaceid": "10d4ea44-d470-4139-96d0-4c880d036717", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6fe29327-b1fc-40ca-9f06-6bc80a449526", "address": "fa:16:3e:d2:b1:83", "network": {"id": "9062a672-c072-445e-80c6-de0be474454d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1899708883", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.65", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe29327-b1", "ovs_interfaceid": "6fe29327-b1fc-40ca-9f06-6bc80a449526", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "address": "fa:16:3e:6f:ef:7e", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2296c391-a8", "ovs_interfaceid": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.374019] env[61824]: DEBUG oslo_vmware.api [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275883, 'name': PowerOffVM_Task, 'duration_secs': 0.243714} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.374442] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.374587] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.374865] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2c8a8c7-3eac-40ca-aced-80fef15c1425 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.483296] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.483522] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.483703] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleting the datastore file [datastore1] 957ae086-5bb4-44f6-b287-5fbcd9742d12 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.484023] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2afd7813-6b4c-47ef-ac1b-62ebecef56e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.495254] env[61824]: DEBUG oslo_vmware.api [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for the task: (returnval){ [ 955.495254] env[61824]: value = "task-1275885" [ 955.495254] env[61824]: _type = "Task" [ 955.495254] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.512100] env[61824]: DEBUG oslo_vmware.api [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275885, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.603265] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6470e2aa-d7d9-4784-839d-ef5601a1b695 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.269s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.609997] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "82928174-1761-4296-bebe-7989c088d542" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.610562] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "82928174-1761-4296-bebe-7989c088d542" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.754404] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Releasing lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.754404] env[61824]: DEBUG nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Instance network_info: |[{"id": "10d4ea44-d470-4139-96d0-4c880d036717", "address": "fa:16:3e:65:a5:7f", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10d4ea44-d4", "ovs_interfaceid": "10d4ea44-d470-4139-96d0-4c880d036717", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6fe29327-b1fc-40ca-9f06-6bc80a449526", "address": "fa:16:3e:d2:b1:83", "network": {"id": "9062a672-c072-445e-80c6-de0be474454d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1899708883", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.65", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe29327-b1", "ovs_interfaceid": "6fe29327-b1fc-40ca-9f06-6bc80a449526", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "address": "fa:16:3e:6f:ef:7e", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2296c391-a8", "ovs_interfaceid": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 955.754404] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] Acquired lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.754404] env[61824]: DEBUG nova.network.neutron [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Refreshing network info cache for port 2296c391-a8f5-4322-ac9f-7059d9ccb541 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 955.756197] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:a5:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10d4ea44-d470-4139-96d0-4c880d036717', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:b1:83', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ca16b6db-4f74-424c-9d36-925ad82cbdd2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fe29327-b1fc-40ca-9f06-6bc80a449526', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:ef:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2bf99f85-3a5c-47c6-a603-e215be6ab0bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2296c391-a8f5-4322-ac9f-7059d9ccb541', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 955.769675] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Creating folder: Project (773a259f132a42479464e120df314bb7). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 955.769675] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed1b547c-4005-4b8e-8540-987565f876eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.781975] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Created folder: Project (773a259f132a42479464e120df314bb7) in parent group-v274074. [ 955.782349] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Creating folder: Instances. Parent ref: group-v274202. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 955.782667] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2291f9cd-7930-46d5-95a8-286df0e9e3bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.794245] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Created folder: Instances in parent group-v274202. [ 955.794802] env[61824]: DEBUG oslo.service.loopingcall [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.794802] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 955.795179] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8024c5b1-c113-4072-98cc-2f21ae94d620 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.826461] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 955.826461] env[61824]: value = "task-1275888" [ 955.826461] env[61824]: _type = "Task" [ 955.826461] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.836142] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275888, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.855755] env[61824]: DEBUG nova.network.neutron [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Successfully updated port: 7832e699-039f-4690-9e62-cdb49ebee107 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.941480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.941480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.941480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.941480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.941480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.946708] env[61824]: INFO nova.compute.manager [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Terminating instance [ 955.950061] env[61824]: DEBUG nova.compute.manager [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.950818] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.951945] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c2192d-b704-434f-9433-3d18a5210549 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.962606] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.962763] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8baf9cc-282e-4821-8577-02588bbaf5f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.971332] env[61824]: DEBUG oslo_vmware.api [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 955.971332] env[61824]: value = "task-1275889" [ 955.971332] env[61824]: _type = "Task" [ 955.971332] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.980940] env[61824]: DEBUG oslo_vmware.api [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275889, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.006209] env[61824]: DEBUG oslo_vmware.api [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Task: {'id': task-1275885, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.206376} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.010019] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.010019] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.010019] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.010019] env[61824]: INFO nova.compute.manager [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Took 1.17 seconds to destroy the instance on the hypervisor. [ 956.010019] env[61824]: DEBUG oslo.service.loopingcall [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.010443] env[61824]: DEBUG nova.compute.manager [-] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.010587] env[61824]: DEBUG nova.network.neutron [-] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 956.115054] env[61824]: DEBUG nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 956.166228] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40605fad-a3ad-4cd1-bdb3-27ac55ab8c77 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.176311] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecde4705-d073-41f0-bb49-7603deebbbbb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.214964] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d4afb1-ced8-4976-9795-ab92796c54b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.223907] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8112f49a-0a3d-46ec-8c00-cfba8bcdb93b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.239290] env[61824]: DEBUG nova.compute.provider_tree [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.251534] env[61824]: DEBUG nova.compute.manager [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Received event network-vif-plugged-7832e699-039f-4690-9e62-cdb49ebee107 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.251534] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] Acquiring lock "44ec5c22-d080-45d1-aa21-080af4c317c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.251682] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.255019] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.255019] env[61824]: DEBUG nova.compute.manager [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] No waiting events found dispatching network-vif-plugged-7832e699-039f-4690-9e62-cdb49ebee107 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.255019] env[61824]: WARNING nova.compute.manager [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Received unexpected event network-vif-plugged-7832e699-039f-4690-9e62-cdb49ebee107 for instance with vm_state building and task_state spawning. [ 956.255019] env[61824]: DEBUG nova.compute.manager [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Received event network-changed-7832e699-039f-4690-9e62-cdb49ebee107 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.255019] env[61824]: DEBUG nova.compute.manager [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Refreshing instance network info cache due to event network-changed-7832e699-039f-4690-9e62-cdb49ebee107. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 956.255019] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] Acquiring lock "refresh_cache-44ec5c22-d080-45d1-aa21-080af4c317c5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.255019] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] Acquired lock "refresh_cache-44ec5c22-d080-45d1-aa21-080af4c317c5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.255019] env[61824]: DEBUG nova.network.neutron [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Refreshing network info cache for port 7832e699-039f-4690-9e62-cdb49ebee107 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 956.335787] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275888, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.358438] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-44ec5c22-d080-45d1-aa21-080af4c317c5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.481204] env[61824]: DEBUG oslo_vmware.api [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275889, 'name': PowerOffVM_Task, 'duration_secs': 0.228082} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.481973] env[61824]: DEBUG nova.network.neutron [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Updated VIF entry in instance network info cache for port 2296c391-a8f5-4322-ac9f-7059d9ccb541. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 956.482433] env[61824]: DEBUG nova.network.neutron [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Updating instance_info_cache with network_info: [{"id": "10d4ea44-d470-4139-96d0-4c880d036717", "address": "fa:16:3e:65:a5:7f", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10d4ea44-d4", "ovs_interfaceid": "10d4ea44-d470-4139-96d0-4c880d036717", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6fe29327-b1fc-40ca-9f06-6bc80a449526", "address": "fa:16:3e:d2:b1:83", "network": {"id": "9062a672-c072-445e-80c6-de0be474454d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1899708883", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.65", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ca16b6db-4f74-424c-9d36-925ad82cbdd2", "external-id": "nsx-vlan-transportzone-200", "segmentation_id": 200, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe29327-b1", "ovs_interfaceid": "6fe29327-b1fc-40ca-9f06-6bc80a449526", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "address": "fa:16:3e:6f:ef:7e", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2296c391-a8", "ovs_interfaceid": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.483640] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.483819] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.484285] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48402951-3148-4424-a3ec-e372b80b3945 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.639219] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.640664] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.640938] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.641140] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Deleting the datastore file [datastore2] 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.641315] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50791496-2ef4-410e-9fd2-724f6f360822 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.649181] env[61824]: DEBUG oslo_vmware.api [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for the task: (returnval){ [ 956.649181] env[61824]: value = "task-1275891" [ 956.649181] env[61824]: _type = "Task" [ 956.649181] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.657569] env[61824]: DEBUG oslo_vmware.api [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275891, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.742211] env[61824]: DEBUG nova.scheduler.client.report [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.810123] env[61824]: DEBUG nova.network.neutron [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.839096] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275888, 'name': CreateVM_Task, 'duration_secs': 0.531996} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.839318] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 956.840261] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.840739] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.841128] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 956.841431] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a773c609-3348-4cce-9a4f-e7bf93526d38 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.845994] env[61824]: DEBUG nova.network.neutron [-] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.847921] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 956.847921] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]529fd09d-3674-f5f0-cc45-6832bcad9422" [ 956.847921] env[61824]: _type = "Task" [ 956.847921] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.859161] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529fd09d-3674-f5f0-cc45-6832bcad9422, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.985049] env[61824]: DEBUG oslo_concurrency.lockutils [req-8d2b4a7d-5ae7-42f2-a344-d6c9a34bcdd6 req-8a3a49dd-98d2-4065-a6aa-0952d78aedbf service nova] Releasing lock "refresh_cache-e500f768-3ec8-48c7-a947-784e8665af15" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.999470] env[61824]: DEBUG nova.network.neutron [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.035245] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.035458] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.035644] env[61824]: DEBUG nova.network.neutron [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.160281] env[61824]: DEBUG oslo_vmware.api [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Task: {'id': task-1275891, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153256} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.160578] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.160770] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.160979] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.161198] env[61824]: INFO nova.compute.manager [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Took 1.21 seconds to destroy the instance on the hypervisor. [ 957.161447] env[61824]: DEBUG oslo.service.loopingcall [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.161638] env[61824]: DEBUG nova.compute.manager [-] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.161731] env[61824]: DEBUG nova.network.neutron [-] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 957.247866] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.248445] env[61824]: DEBUG nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.251121] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.612s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.253016] env[61824]: INFO nova.compute.claims [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.349207] env[61824]: INFO nova.compute.manager [-] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Took 1.34 seconds to deallocate network for instance. [ 957.365486] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529fd09d-3674-f5f0-cc45-6832bcad9422, 'name': SearchDatastore_Task, 'duration_secs': 0.009547} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.365832] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.366133] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 957.366409] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.366604] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.367577] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 957.367577] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56017c4c-faaa-46e0-b955-fa5839846032 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.379684] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 957.379892] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 957.380761] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c424277-9192-4b03-b261-69488cc886aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.386543] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 957.386543] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5247c571-1acb-7b2e-ae22-a85c800bfc68" [ 957.386543] env[61824]: _type = "Task" [ 957.386543] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.395643] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5247c571-1acb-7b2e-ae22-a85c800bfc68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.502337] env[61824]: DEBUG oslo_concurrency.lockutils [req-a4c16432-c9d2-459d-b8dc-2d7a5592ef40 req-cbe78689-e045-411d-b9de-68280ecf2575 service nova] Releasing lock "refresh_cache-44ec5c22-d080-45d1-aa21-080af4c317c5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.502743] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-44ec5c22-d080-45d1-aa21-080af4c317c5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.502912] env[61824]: DEBUG nova.network.neutron [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.757205] env[61824]: DEBUG nova.compute.utils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.761697] env[61824]: DEBUG nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.762050] env[61824]: DEBUG nova.network.neutron [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 957.822666] env[61824]: DEBUG nova.policy [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1eb545ac2cf4f0d9ccddb040b17372b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e756c6817698464a9da391d4dd364946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.854029] env[61824]: DEBUG nova.network.neutron [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45dc4201-9f", "ovs_interfaceid": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.861520] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.902757] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5247c571-1acb-7b2e-ae22-a85c800bfc68, 'name': SearchDatastore_Task, 'duration_secs': 0.010986} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.903840] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7be886b1-08bd-45f7-9491-bfa3c949e708 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.909479] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 957.909479] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52396cd2-63c4-6105-a773-2bc5310114a2" [ 957.909479] env[61824]: _type = "Task" [ 957.909479] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.918808] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52396cd2-63c4-6105-a773-2bc5310114a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.090663] env[61824]: DEBUG nova.network.neutron [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 958.155599] env[61824]: DEBUG nova.network.neutron [-] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.262743] env[61824]: DEBUG nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.357534] env[61824]: DEBUG nova.compute.manager [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Received event network-vif-deleted-92b95e81-3307-478e-91c4-bfe7dd132b7d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.357749] env[61824]: DEBUG nova.compute.manager [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received event network-vif-plugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.357993] env[61824]: DEBUG oslo_concurrency.lockutils [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.359147] env[61824]: DEBUG oslo_concurrency.lockutils [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.359147] env[61824]: DEBUG oslo_concurrency.lockutils [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.359492] env[61824]: DEBUG nova.compute.manager [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] No waiting events found dispatching network-vif-plugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.359492] env[61824]: WARNING nova.compute.manager [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received unexpected event network-vif-plugged-45dc4201-9fb0-4060-ba24-f376e69a6e45 for instance with vm_state shelved_offloaded and task_state spawning. [ 958.359756] env[61824]: DEBUG nova.compute.manager [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received event network-changed-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.359997] env[61824]: DEBUG nova.compute.manager [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Refreshing instance network info cache due to event network-changed-45dc4201-9fb0-4060-ba24-f376e69a6e45. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.360230] env[61824]: DEBUG oslo_concurrency.lockutils [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] Acquiring lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.360887] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.365908] env[61824]: DEBUG oslo_concurrency.lockutils [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] Acquired lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.366236] env[61824]: DEBUG nova.network.neutron [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Refreshing network info cache for port 45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 958.389649] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='ea542d49a00460e320932ff1b1365ed7',container_format='bare',created_at=2024-10-12T10:59:29Z,direct_url=,disk_format='vmdk',id=0da722b8-2f9f-4a79-ada2-0ce4f4dd247e,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1631666102-shelved',owner='dc352363e7f94dc6b7a99db8aca565a6',properties=ImageMetaProps,protected=,size=31665152,status='active',tags=,updated_at=2024-10-12T10:59:45Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.389889] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.390181] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.390708] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.390708] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.390708] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.390864] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.390901] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.391215] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.391447] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.391682] env[61824]: DEBUG nova.virt.hardware [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.392794] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a794108c-064d-459a-b9b6-705680ac0aa3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.405467] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b1ab01-9a5c-4875-95c4-b3b0c0b3bb41 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.426228] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:8d:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '55764410-260e-4339-a020-6b30995584bf', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45dc4201-9fb0-4060-ba24-f376e69a6e45', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.433631] env[61824]: DEBUG oslo.service.loopingcall [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.440193] env[61824]: DEBUG nova.network.neutron [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Successfully created port: 22e2de70-07bb-4d94-8624-d6082438907b {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.442222] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.442554] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52396cd2-63c4-6105-a773-2bc5310114a2, 'name': SearchDatastore_Task, 'duration_secs': 0.010185} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.443037] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38d83607-f401-4ee1-a6eb-900f579cc329 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.458842] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.459172] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] e500f768-3ec8-48c7-a947-784e8665af15/e500f768-3ec8-48c7-a947-784e8665af15.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 958.460419] env[61824]: DEBUG nova.network.neutron [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Updating instance_info_cache with network_info: [{"id": "7832e699-039f-4690-9e62-cdb49ebee107", "address": "fa:16:3e:38:80:01", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7832e699-03", "ovs_interfaceid": "7832e699-039f-4690-9e62-cdb49ebee107", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.463435] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56f5c5b5-cbd6-4fe0-9302-bc3e3c9b8e05 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.465768] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9faecd72-22cf-48c6-9985-fc48a52c4232 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.474746] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0913f7c2-b868-407a-a09b-9cf2071c587b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.478072] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.478072] env[61824]: value = "task-1275893" [ 958.478072] env[61824]: _type = "Task" [ 958.478072] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.480113] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 958.480113] env[61824]: value = "task-1275892" [ 958.480113] env[61824]: _type = "Task" [ 958.480113] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.518773] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0fecc0-47e8-497a-a7a1-416678163507 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.525458] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275893, 'name': CreateVM_Task} progress is 15%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.531177] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275892, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.534705] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7361b065-cf59-46e1-aee7-d2dcedb790cb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.553547] env[61824]: DEBUG nova.compute.provider_tree [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.658527] env[61824]: INFO nova.compute.manager [-] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Took 1.50 seconds to deallocate network for instance. [ 958.969487] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-44ec5c22-d080-45d1-aa21-080af4c317c5" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.969844] env[61824]: DEBUG nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Instance network_info: |[{"id": "7832e699-039f-4690-9e62-cdb49ebee107", "address": "fa:16:3e:38:80:01", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7832e699-03", "ovs_interfaceid": "7832e699-039f-4690-9e62-cdb49ebee107", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.970291] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:80:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7832e699-039f-4690-9e62-cdb49ebee107', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.978060] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating folder: Project (629276280a234e7a8b810e6e74779fff). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.978649] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-33c7f5b4-8425-45a8-83ba-699525f8f741 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.991741] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275893, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.995462] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275892, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457593} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.996815] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] e500f768-3ec8-48c7-a947-784e8665af15/e500f768-3ec8-48c7-a947-784e8665af15.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 958.997080] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 958.997359] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created folder: Project (629276280a234e7a8b810e6e74779fff) in parent group-v274074. [ 958.997528] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating folder: Instances. Parent ref: group-v274206. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 958.997749] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4fc994a-80d2-44e2-89ba-b6bc264a4631 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.999849] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c905f569-4d08-4c63-b78c-ee768686b7f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.009710] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 959.009710] env[61824]: value = "task-1275896" [ 959.009710] env[61824]: _type = "Task" [ 959.009710] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.014332] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created folder: Instances in parent group-v274206. [ 959.014581] env[61824]: DEBUG oslo.service.loopingcall [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.015138] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 959.015360] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-81dd6ee9-c471-4675-acd4-8bb02f9172b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.033338] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275896, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.043925] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 959.043925] env[61824]: value = "task-1275897" [ 959.043925] env[61824]: _type = "Task" [ 959.043925] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.052510] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275897, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.056498] env[61824]: DEBUG nova.scheduler.client.report [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.119873] env[61824]: DEBUG nova.network.neutron [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updated VIF entry in instance network info cache for port 45dc4201-9fb0-4060-ba24-f376e69a6e45. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 959.120555] env[61824]: DEBUG nova.network.neutron [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45dc4201-9f", "ovs_interfaceid": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.166120] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.275749] env[61824]: DEBUG nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.299333] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.299596] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.299758] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.299942] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.300114] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.300268] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.300477] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.300641] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.300813] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.300980] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.301176] env[61824]: DEBUG nova.virt.hardware [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.302066] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e188938f-05d3-4da6-b9dc-08a9397475f4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.311550] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183a85ce-92d9-4990-9d0b-b990e708f71a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.491541] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275893, 'name': CreateVM_Task, 'duration_secs': 0.57488} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.491724] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.492431] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.492607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.492990] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.493274] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b7dba42-a7bb-497c-a641-d4408b542658 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.498437] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 959.498437] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5245ea84-5d4f-9624-a1b2-1b612bbf4ea0" [ 959.498437] env[61824]: _type = "Task" [ 959.498437] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.506430] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5245ea84-5d4f-9624-a1b2-1b612bbf4ea0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.519149] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275896, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06868} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.519410] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 959.520224] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50bc700-e96a-48fb-839e-f0a7562b4612 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.548098] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] e500f768-3ec8-48c7-a947-784e8665af15/e500f768-3ec8-48c7-a947-784e8665af15.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 959.548452] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b51ad51-4197-45f2-9d65-092d77f17bcb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.567566] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.568095] env[61824]: DEBUG nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 959.570763] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.711s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.571013] env[61824]: DEBUG nova.objects.instance [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lazy-loading 'resources' on Instance uuid 957ae086-5bb4-44f6-b287-5fbcd9742d12 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.578221] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275897, 'name': CreateVM_Task, 'duration_secs': 0.326173} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.579493] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 959.580061] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 959.580061] env[61824]: value = "task-1275898" [ 959.580061] env[61824]: _type = "Task" [ 959.580061] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.580687] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.580965] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.581311] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.581643] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1911209f-f751-4870-a579-2c006102e567 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.593348] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275898, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.594316] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 959.594316] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5277ece5-4efb-1f4c-3711-8e5bf38f5364" [ 959.594316] env[61824]: _type = "Task" [ 959.594316] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.603544] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5277ece5-4efb-1f4c-3711-8e5bf38f5364, 'name': SearchDatastore_Task, 'duration_secs': 0.009833} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.603820] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.604067] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.604304] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.604451] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.604631] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.604899] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d5f6483d-7916-4c2c-badc-653d68a92a2a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.614564] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.614761] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.615639] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00eef8e6-7c57-48ce-aeed-baec787fa5be {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.622218] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 959.622218] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52867b9b-044c-cdfb-102c-f42f3c8fba9e" [ 959.622218] env[61824]: _type = "Task" [ 959.622218] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.626609] env[61824]: DEBUG oslo_concurrency.lockutils [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] Releasing lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.627123] env[61824]: DEBUG nova.compute.manager [req-e7aeb691-209c-48f2-b858-42b0d5c54691 req-46cd39c8-8ecd-43e4-9694-0e5dccfee3e5 service nova] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Received event network-vif-deleted-41458d47-a6e3-44c2-a378-9968713e2739 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.633308] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52867b9b-044c-cdfb-102c-f42f3c8fba9e, 'name': SearchDatastore_Task, 'duration_secs': 0.008768} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.633999] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7396fc84-82ae-48ef-afb9-65e836d39f55 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.639499] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 959.639499] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]528a9322-34fd-0aec-fdff-6fcdccd59cf5" [ 959.639499] env[61824]: _type = "Task" [ 959.639499] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.650210] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528a9322-34fd-0aec-fdff-6fcdccd59cf5, 'name': SearchDatastore_Task, 'duration_secs': 0.008305} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.650455] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.650709] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 44ec5c22-d080-45d1-aa21-080af4c317c5/44ec5c22-d080-45d1-aa21-080af4c317c5.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.650972] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47b5c675-4266-40a3-8b3a-70b70b9d194b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.659642] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 959.659642] env[61824]: value = "task-1275899" [ 959.659642] env[61824]: _type = "Task" [ 959.659642] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.667800] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275899, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.013501] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.014506] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Processing image 0da722b8-2f9f-4a79-ada2-0ce4f4dd247e {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.014506] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.014656] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.014907] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.015292] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4125800e-b366-4650-bc0d-1afacac71645 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.031123] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.031362] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.032300] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-190068ec-ea31-4699-bfea-55e68dcbab18 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.040409] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 960.040409] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52588c3f-a79d-0f3b-6c73-9e9806652829" [ 960.040409] env[61824]: _type = "Task" [ 960.040409] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.053579] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52588c3f-a79d-0f3b-6c73-9e9806652829, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.070748] env[61824]: DEBUG nova.network.neutron [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Successfully updated port: 22e2de70-07bb-4d94-8624-d6082438907b {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.076022] env[61824]: DEBUG nova.compute.utils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.076022] env[61824]: DEBUG nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.076636] env[61824]: DEBUG nova.network.neutron [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 960.096184] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275898, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.127308] env[61824]: DEBUG nova.policy [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e8f7abae6ee4e9690831b9941c1d35b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd358ecdc9034719bdb077c1e7c8cb35', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.171343] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275899, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507894} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.174366] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 44ec5c22-d080-45d1-aa21-080af4c317c5/44ec5c22-d080-45d1-aa21-080af4c317c5.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.174612] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.175202] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d30917be-ee0e-41b4-9fc4-a73d73f53365 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.184402] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 960.184402] env[61824]: value = "task-1275900" [ 960.184402] env[61824]: _type = "Task" [ 960.184402] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.198857] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275900, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.256844] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aba3a36-c950-4c90-9251-98f8edd29c6d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.264699] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e37b748-5dc4-479d-97d4-7caa0409e9ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.296723] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04bfc22a-6928-4350-8e92-9ac016e118c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.304420] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2c60d1-c288-40a7-a76f-2f4f493c6b56 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.317640] env[61824]: DEBUG nova.compute.provider_tree [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.383666] env[61824]: DEBUG nova.compute.manager [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Received event network-vif-plugged-22e2de70-07bb-4d94-8624-d6082438907b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.383901] env[61824]: DEBUG oslo_concurrency.lockutils [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] Acquiring lock "3256627d-35c8-4b75-b4a8-390eda0a150e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.384127] env[61824]: DEBUG oslo_concurrency.lockutils [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.384298] env[61824]: DEBUG oslo_concurrency.lockutils [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.384509] env[61824]: DEBUG nova.compute.manager [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] No waiting events found dispatching network-vif-plugged-22e2de70-07bb-4d94-8624-d6082438907b {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 960.384622] env[61824]: WARNING nova.compute.manager [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Received unexpected event network-vif-plugged-22e2de70-07bb-4d94-8624-d6082438907b for instance with vm_state building and task_state spawning. [ 960.384784] env[61824]: DEBUG nova.compute.manager [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Received event network-changed-22e2de70-07bb-4d94-8624-d6082438907b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 960.385037] env[61824]: DEBUG nova.compute.manager [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Refreshing instance network info cache due to event network-changed-22e2de70-07bb-4d94-8624-d6082438907b. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 960.385817] env[61824]: DEBUG oslo_concurrency.lockutils [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] Acquiring lock "refresh_cache-3256627d-35c8-4b75-b4a8-390eda0a150e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.386357] env[61824]: DEBUG oslo_concurrency.lockutils [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] Acquired lock "refresh_cache-3256627d-35c8-4b75-b4a8-390eda0a150e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.386357] env[61824]: DEBUG nova.network.neutron [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Refreshing network info cache for port 22e2de70-07bb-4d94-8624-d6082438907b {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 960.412395] env[61824]: DEBUG nova.network.neutron [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Successfully created port: 44ea9f2e-3382-431f-b83a-590af6c7f511 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.551299] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Preparing fetch location {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 960.551566] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Fetch image to [datastore1] OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f/OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f.vmdk {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 960.551751] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Downloading stream optimized image 0da722b8-2f9f-4a79-ada2-0ce4f4dd247e to [datastore1] OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f/OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f.vmdk on the data store datastore1 as vApp {{(pid=61824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 960.551925] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Downloading image file data 0da722b8-2f9f-4a79-ada2-0ce4f4dd247e to the ESX as VM named 'OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f' {{(pid=61824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 960.590215] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "refresh_cache-3256627d-35c8-4b75-b4a8-390eda0a150e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.590752] env[61824]: DEBUG nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 960.603654] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275898, 'name': ReconfigVM_Task, 'duration_secs': 0.646494} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.603934] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Reconfigured VM instance instance-0000005a to attach disk [datastore1] e500f768-3ec8-48c7-a947-784e8665af15/e500f768-3ec8-48c7-a947-784e8665af15.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 960.604585] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7143da97-e1a8-4f0c-b802-2b59cc206171 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.613430] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 960.613430] env[61824]: value = "task-1275901" [ 960.613430] env[61824]: _type = "Task" [ 960.613430] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.624602] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275901, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.650800] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 960.650800] env[61824]: value = "resgroup-9" [ 960.650800] env[61824]: _type = "ResourcePool" [ 960.650800] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 960.651378] env[61824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b6b16c88-1e35-4a39-ab34-2a21a44588c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.675016] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lease: (returnval){ [ 960.675016] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a7084f-0001-151c-e5e0-30fc3cfb46d3" [ 960.675016] env[61824]: _type = "HttpNfcLease" [ 960.675016] env[61824]: } obtained for vApp import into resource pool (val){ [ 960.675016] env[61824]: value = "resgroup-9" [ 960.675016] env[61824]: _type = "ResourcePool" [ 960.675016] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 960.675399] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the lease: (returnval){ [ 960.675399] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a7084f-0001-151c-e5e0-30fc3cfb46d3" [ 960.675399] env[61824]: _type = "HttpNfcLease" [ 960.675399] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 960.682474] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.682474] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a7084f-0001-151c-e5e0-30fc3cfb46d3" [ 960.682474] env[61824]: _type = "HttpNfcLease" [ 960.682474] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.685420] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 960.688394] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 960.695156] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275900, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06905} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.695432] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.696237] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e370e9a6-889b-4a98-8ee1-6bca326d14a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.720542] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 44ec5c22-d080-45d1-aa21-080af4c317c5/44ec5c22-d080-45d1-aa21-080af4c317c5.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.721465] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1f4c27b-c865-44e5-b794-23bd3c218afb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.745294] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 960.745294] env[61824]: value = "task-1275903" [ 960.745294] env[61824]: _type = "Task" [ 960.745294] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.755153] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275903, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.820430] env[61824]: DEBUG nova.scheduler.client.report [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.950323] env[61824]: DEBUG nova.network.neutron [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.040701] env[61824]: DEBUG nova.network.neutron [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.122800] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275901, 'name': Rename_Task, 'duration_secs': 0.168542} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.122957] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.123099] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40fcdfa7-4e8e-4af1-b1f1-93028da21e47 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.131042] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 961.131042] env[61824]: value = "task-1275904" [ 961.131042] env[61824]: _type = "Task" [ 961.131042] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.141561] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275904, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.184110] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 961.184110] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a7084f-0001-151c-e5e0-30fc3cfb46d3" [ 961.184110] env[61824]: _type = "HttpNfcLease" [ 961.184110] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 961.195677] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 961.195856] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 961.256842] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275903, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.325874] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.328958] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.163s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.329492] env[61824]: DEBUG nova.objects.instance [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lazy-loading 'resources' on Instance uuid 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.347063] env[61824]: INFO nova.scheduler.client.report [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Deleted allocations for instance 957ae086-5bb4-44f6-b287-5fbcd9742d12 [ 961.543648] env[61824]: DEBUG oslo_concurrency.lockutils [req-f1634145-4d6a-4e16-9d61-f4b8fe08161b req-39c009b9-0124-4909-82ed-12c3b19abef0 service nova] Releasing lock "refresh_cache-3256627d-35c8-4b75-b4a8-390eda0a150e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.544011] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "refresh_cache-3256627d-35c8-4b75-b4a8-390eda0a150e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.544195] env[61824]: DEBUG nova.network.neutron [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.604933] env[61824]: DEBUG nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 961.633152] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.633470] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.633635] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.633824] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.633977] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.634345] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.634574] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.634745] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.634927] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.635106] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.635321] env[61824]: DEBUG nova.virt.hardware [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.636256] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d74e3a-86bb-44cf-8edf-28aa3a7ac008 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.650486] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-343220b5-0222-4641-9167-0ddc777b6afa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.654384] env[61824]: DEBUG oslo_vmware.api [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275904, 'name': PowerOnVM_Task, 'duration_secs': 0.490516} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.654618] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 961.654818] env[61824]: INFO nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Took 15.37 seconds to spawn the instance on the hypervisor. [ 961.654998] env[61824]: DEBUG nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 961.656155] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc59afa-5b44-463b-8db5-55a9dda6ef8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.683825] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 961.683825] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a7084f-0001-151c-e5e0-30fc3cfb46d3" [ 961.683825] env[61824]: _type = "HttpNfcLease" [ 961.683825] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 961.684239] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 961.684239] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a7084f-0001-151c-e5e0-30fc3cfb46d3" [ 961.684239] env[61824]: _type = "HttpNfcLease" [ 961.684239] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 961.684984] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d5074a-e4dd-430c-82bb-6c882e99bfae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.692875] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b422b-81cf-f206-3c26-24ee8465b8fb/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 961.693124] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating HTTP connection to write to file with size = 31665152 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b422b-81cf-f206-3c26-24ee8465b8fb/disk-0.vmdk. {{(pid=61824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 961.765035] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-63f24cb8-9f89-41d9-8854-777c84c4b647 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.766177] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275903, 'name': ReconfigVM_Task, 'duration_secs': 0.768772} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.767628] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 44ec5c22-d080-45d1-aa21-080af4c317c5/44ec5c22-d080-45d1-aa21-080af4c317c5.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.768890] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-116ec4b5-ce45-462b-b183-7e9d258e1be2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.777870] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 961.777870] env[61824]: value = "task-1275905" [ 961.777870] env[61824]: _type = "Task" [ 961.777870] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.788066] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275905, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.857916] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d4045912-a493-415e-9f3f-60254e40815b tempest-ServerDiskConfigTestJSON-1964998631 tempest-ServerDiskConfigTestJSON-1964998631-project-member] Lock "957ae086-5bb4-44f6-b287-5fbcd9742d12" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.025s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.993555] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f521bd-6197-40e6-ba39-d405ef83c1ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.002576] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e864f938-6f38-466d-9168-39d5836cfa58 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.047483] env[61824]: DEBUG nova.network.neutron [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Successfully updated port: 44ea9f2e-3382-431f-b83a-590af6c7f511 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.049349] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6155ac62-b84e-4593-8589-d0a54deccd92 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.066157] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9baae666-ebcf-4718-940a-7e96035c45e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.081840] env[61824]: DEBUG nova.compute.provider_tree [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.122887] env[61824]: DEBUG nova.network.neutron [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 962.189705] env[61824]: INFO nova.compute.manager [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Took 24.66 seconds to build instance. [ 962.291304] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275905, 'name': Rename_Task, 'duration_secs': 0.139087} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.291604] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 962.291859] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2647305d-63fe-41f1-9630-134e58880745 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.302306] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 962.302306] env[61824]: value = "task-1275906" [ 962.302306] env[61824]: _type = "Task" [ 962.302306] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.313910] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.397462] env[61824]: DEBUG nova.network.neutron [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Updating instance_info_cache with network_info: [{"id": "22e2de70-07bb-4d94-8624-d6082438907b", "address": "fa:16:3e:cc:f0:2e", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22e2de70-07", "ovs_interfaceid": "22e2de70-07bb-4d94-8624-d6082438907b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.485790] env[61824]: DEBUG nova.compute.manager [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Received event network-vif-plugged-44ea9f2e-3382-431f-b83a-590af6c7f511 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.488016] env[61824]: DEBUG oslo_concurrency.lockutils [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] Acquiring lock "82928174-1761-4296-bebe-7989c088d542-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.488016] env[61824]: DEBUG oslo_concurrency.lockutils [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] Lock "82928174-1761-4296-bebe-7989c088d542-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.488016] env[61824]: DEBUG oslo_concurrency.lockutils [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] Lock "82928174-1761-4296-bebe-7989c088d542-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.488016] env[61824]: DEBUG nova.compute.manager [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] No waiting events found dispatching network-vif-plugged-44ea9f2e-3382-431f-b83a-590af6c7f511 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 962.488016] env[61824]: WARNING nova.compute.manager [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Received unexpected event network-vif-plugged-44ea9f2e-3382-431f-b83a-590af6c7f511 for instance with vm_state building and task_state spawning. [ 962.488016] env[61824]: DEBUG nova.compute.manager [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Received event network-changed-44ea9f2e-3382-431f-b83a-590af6c7f511 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.488016] env[61824]: DEBUG nova.compute.manager [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Refreshing instance network info cache due to event network-changed-44ea9f2e-3382-431f-b83a-590af6c7f511. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 962.488016] env[61824]: DEBUG oslo_concurrency.lockutils [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] Acquiring lock "refresh_cache-82928174-1761-4296-bebe-7989c088d542" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.488016] env[61824]: DEBUG oslo_concurrency.lockutils [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] Acquired lock "refresh_cache-82928174-1761-4296-bebe-7989c088d542" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.488016] env[61824]: DEBUG nova.network.neutron [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Refreshing network info cache for port 44ea9f2e-3382-431f-b83a-590af6c7f511 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.557941] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "refresh_cache-82928174-1761-4296-bebe-7989c088d542" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.587064] env[61824]: DEBUG nova.scheduler.client.report [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.693513] env[61824]: DEBUG oslo_concurrency.lockutils [None req-db125d14-b267-45e8-82c4-1d4c51be3222 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "e500f768-3ec8-48c7-a947-784e8665af15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.179s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.814241] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275906, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.846556] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "e500f768-3ec8-48c7-a947-784e8665af15" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.846969] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "e500f768-3ec8-48c7-a947-784e8665af15" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.847211] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "e500f768-3ec8-48c7-a947-784e8665af15-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.847431] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.847674] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "e500f768-3ec8-48c7-a947-784e8665af15-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.850374] env[61824]: INFO nova.compute.manager [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Terminating instance [ 962.852462] env[61824]: DEBUG nova.compute.manager [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 962.852678] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 962.853620] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3c58c4-6072-44c0-8d47-6a4fe391f93d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.863845] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.864282] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6a3a39d-e2f4-4910-96ee-e1e5a4353390 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.872657] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 962.872657] env[61824]: value = "task-1275907" [ 962.872657] env[61824]: _type = "Task" [ 962.872657] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.883437] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.901436] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "refresh_cache-3256627d-35c8-4b75-b4a8-390eda0a150e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.901948] env[61824]: DEBUG nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Instance network_info: |[{"id": "22e2de70-07bb-4d94-8624-d6082438907b", "address": "fa:16:3e:cc:f0:2e", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22e2de70-07", "ovs_interfaceid": "22e2de70-07bb-4d94-8624-d6082438907b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.902475] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:f0:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22e2de70-07bb-4d94-8624-d6082438907b', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.912432] env[61824]: DEBUG oslo.service.loopingcall [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.916264] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 962.916373] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5f97b40-7f8c-4f0a-8df1-d47b88e38b67 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.936299] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Completed reading data from the image iterator. {{(pid=61824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 962.936558] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b422b-81cf-f206-3c26-24ee8465b8fb/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 962.937542] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3525e2c2-f757-4a08-ab44-ae8a61bb4da5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.946304] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b422b-81cf-f206-3c26-24ee8465b8fb/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 962.946484] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b422b-81cf-f206-3c26-24ee8465b8fb/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 962.947819] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-28a52730-df17-4a46-b875-55524dca0949 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.949329] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.949329] env[61824]: value = "task-1275908" [ 962.949329] env[61824]: _type = "Task" [ 962.949329] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.958345] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275908, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.028027] env[61824]: DEBUG nova.network.neutron [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 963.094148] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.125855] env[61824]: INFO nova.scheduler.client.report [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Deleted allocations for instance 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8 [ 963.304814] env[61824]: DEBUG nova.network.neutron [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.304814] env[61824]: DEBUG oslo_vmware.rw_handles [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b422b-81cf-f206-3c26-24ee8465b8fb/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 963.304814] env[61824]: INFO nova.virt.vmwareapi.images [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Downloaded image file data 0da722b8-2f9f-4a79-ada2-0ce4f4dd247e [ 963.304814] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68d4544-f8a6-4292-9a34-145bf3f97a52 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.304814] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf2b4f28-ef91-4370-b43e-77d2c5a01280 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.304814] env[61824]: INFO nova.virt.vmwareapi.images [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] The imported VM was unregistered [ 963.304814] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Caching image {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 963.304814] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Creating directory with path [datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.304814] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e01ef48d-7e19-4650-842c-c1326c96ea29 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.304814] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Created directory with path [datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.304814] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f/OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f.vmdk to [datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e.vmdk. {{(pid=61824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 963.304814] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d8e68732-d3de-413b-b5de-44ae56c77863 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.304814] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 963.304814] env[61824]: value = "task-1275910" [ 963.304814] env[61824]: _type = "Task" [ 963.304814] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.309229] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275910, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.318541] env[61824]: DEBUG oslo_vmware.api [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275906, 'name': PowerOnVM_Task, 'duration_secs': 0.593462} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.318880] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 963.319115] env[61824]: INFO nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Took 9.57 seconds to spawn the instance on the hypervisor. [ 963.319483] env[61824]: DEBUG nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.320409] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a82926-c621-490e-9ec0-2085348e179a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.332875] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.333037] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquired lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.333182] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Forcefully refreshing network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 963.386882] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275907, 'name': PowerOffVM_Task, 'duration_secs': 0.221359} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.387664] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 963.387823] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 963.388291] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2c4878a-8774-490a-b96c-9958a1a4bf90 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.461040] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275908, 'name': CreateVM_Task, 'duration_secs': 0.37715} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.461284] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 963.462333] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.462333] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.462761] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 963.463214] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6816ce7b-284a-4ad8-9e48-06fa07ccb22c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.469842] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 963.469842] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52ac9b71-6485-4550-f962-3733e9c2f7e8" [ 963.469842] env[61824]: _type = "Task" [ 963.469842] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.480771] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ac9b71-6485-4550-f962-3733e9c2f7e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.560198] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 963.560485] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 963.560675] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Deleting the datastore file [datastore1] e500f768-3ec8-48c7-a947-784e8665af15 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 963.560973] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d2bcaf6-88fe-45e8-9e22-a54b3590f45c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.572253] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 963.572253] env[61824]: value = "task-1275912" [ 963.572253] env[61824]: _type = "Task" [ 963.572253] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.587020] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.635227] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f41e1001-e53f-4640-a77b-b3ccf4507df1 tempest-AttachVolumeNegativeTest-1173229333 tempest-AttachVolumeNegativeTest-1173229333-project-member] Lock "0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.695s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.690671] env[61824]: DEBUG oslo_concurrency.lockutils [req-21a050b1-755c-4958-bde0-aaee204c1cbb req-4de730e1-dd0f-4533-887d-4a2984239209 service nova] Releasing lock "refresh_cache-82928174-1761-4296-bebe-7989c088d542" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.691168] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquired lock "refresh_cache-82928174-1761-4296-bebe-7989c088d542" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.691334] env[61824]: DEBUG nova.network.neutron [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 963.809187] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275910, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.842129] env[61824]: INFO nova.compute.manager [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Took 20.82 seconds to build instance. [ 963.989783] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ac9b71-6485-4550-f962-3733e9c2f7e8, 'name': SearchDatastore_Task, 'duration_secs': 0.018295} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.990498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.990498] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.990701] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.990780] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.990964] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.991266] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-903e4f6f-5b9b-4823-9dc9-8479bb883128 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.006025] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 964.006025] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 964.006025] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcddfd6d-e001-4f6a-9466-c96c24e90e91 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.013024] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 964.013024] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525717f4-0be2-9d01-915e-e4a884529e4b" [ 964.013024] env[61824]: _type = "Task" [ 964.013024] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.026686] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525717f4-0be2-9d01-915e-e4a884529e4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.086102] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.232388] env[61824]: DEBUG nova.network.neutron [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 964.314945] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275910, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.345202] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2dc6b5f8-98b9-4e2c-94b8-7196427f932c tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.328s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.445801] env[61824]: DEBUG nova.network.neutron [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Updating instance_info_cache with network_info: [{"id": "44ea9f2e-3382-431f-b83a-590af6c7f511", "address": "fa:16:3e:74:3a:16", "network": {"id": "aaf225e7-9bce-46be-bc4e-2d3bd93ab681", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1207832912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd358ecdc9034719bdb077c1e7c8cb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c06e3c2-8edb-4cf0-be6b-45dfe059c00b", "external-id": "nsx-vlan-transportzone-264", "segmentation_id": 264, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44ea9f2e-33", "ovs_interfaceid": "44ea9f2e-3382-431f-b83a-590af6c7f511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.529560] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525717f4-0be2-9d01-915e-e4a884529e4b, 'name': SearchDatastore_Task, 'duration_secs': 0.013606} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.530590] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e22e3d1-4cba-41dc-af89-6e8e9a1f3f31 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.538400] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 964.538400] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52320a9d-2aaa-aead-a745-5b97754f6d26" [ 964.538400] env[61824]: _type = "Task" [ 964.538400] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.550690] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52320a9d-2aaa-aead-a745-5b97754f6d26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.588117] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.763992] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [{"id": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "address": "fa:16:3e:a5:5b:14", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa66cb8c8-58", "ovs_interfaceid": "a66cb8c8-5826-4e02-ad96-b2bbea85d23b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.810419] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275910, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.951839] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Releasing lock "refresh_cache-82928174-1761-4296-bebe-7989c088d542" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.952219] env[61824]: DEBUG nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Instance network_info: |[{"id": "44ea9f2e-3382-431f-b83a-590af6c7f511", "address": "fa:16:3e:74:3a:16", "network": {"id": "aaf225e7-9bce-46be-bc4e-2d3bd93ab681", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1207832912-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cd358ecdc9034719bdb077c1e7c8cb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2c06e3c2-8edb-4cf0-be6b-45dfe059c00b", "external-id": "nsx-vlan-transportzone-264", "segmentation_id": 264, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44ea9f2e-33", "ovs_interfaceid": "44ea9f2e-3382-431f-b83a-590af6c7f511", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 964.952686] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:3a:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2c06e3c2-8edb-4cf0-be6b-45dfe059c00b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44ea9f2e-3382-431f-b83a-590af6c7f511', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.962792] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Creating folder: Project (cd358ecdc9034719bdb077c1e7c8cb35). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 964.964165] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a6ee2690-a561-44cc-a08c-1a9c22e77c9b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.979140] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Created folder: Project (cd358ecdc9034719bdb077c1e7c8cb35) in parent group-v274074. [ 964.981024] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Creating folder: Instances. Parent ref: group-v274211. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 964.981024] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d658c25e-b840-4e55-8ff4-0ede5c27c7f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.996161] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Created folder: Instances in parent group-v274211. [ 964.996161] env[61824]: DEBUG oslo.service.loopingcall [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.996161] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82928174-1761-4296-bebe-7989c088d542] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 964.996161] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aaaeb4a4-86de-4f40-890a-0f56db9226e4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.021898] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 965.021898] env[61824]: value = "task-1275916" [ 965.021898] env[61824]: _type = "Task" [ 965.021898] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.033260] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275916, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.052125] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52320a9d-2aaa-aead-a745-5b97754f6d26, 'name': SearchDatastore_Task, 'duration_secs': 0.016903} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.052460] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.052741] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 3256627d-35c8-4b75-b4a8-390eda0a150e/3256627d-35c8-4b75-b4a8-390eda0a150e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 965.053055] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24fbd547-1196-4f09-93d8-6ce4b3c2c22c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.063803] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 965.063803] env[61824]: value = "task-1275917" [ 965.063803] env[61824]: _type = "Task" [ 965.063803] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.075963] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275917, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.087773] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.263215] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "44ec5c22-d080-45d1-aa21-080af4c317c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.263639] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.263959] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "44ec5c22-d080-45d1-aa21-080af4c317c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.264220] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.264494] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.266992] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Releasing lock "refresh_cache-7be79422-c624-4b4a-884f-29b13b9d81d2" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.266992] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updated the network info_cache for instance {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 965.267531] env[61824]: INFO nova.compute.manager [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Terminating instance [ 965.269431] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.270131] env[61824]: DEBUG nova.compute.manager [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 965.270486] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.270708] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.271712] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823d5cf9-420c-4aa4-a448-d899546ec55e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.275261] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.275545] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.276289] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.276503] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.276683] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 965.276866] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 965.285061] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.285434] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b043d249-72c1-4d4a-8475-60a956051479 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.295728] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 965.295728] env[61824]: value = "task-1275918" [ 965.295728] env[61824]: _type = "Task" [ 965.295728] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.312429] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275910, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.316172] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275918, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.378595] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.379121] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.536614] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275916, 'name': CreateVM_Task, 'duration_secs': 0.400401} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.536940] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 82928174-1761-4296-bebe-7989c088d542] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 965.537862] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.538675] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.538876] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 965.539645] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cb01b4c-28f8-4c24-9fb6-a983a364391b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.547179] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 965.547179] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d8744b-dcaa-9d4d-3d22-bacb71a5e142" [ 965.547179] env[61824]: _type = "Task" [ 965.547179] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.558257] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d8744b-dcaa-9d4d-3d22-bacb71a5e142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.577855] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275917, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.590698] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275912, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.781898] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.782161] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.783547] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.783547] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 965.783547] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d5bc75-f5c4-4cf6-9162-e96ba19ab79f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.794749] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912667bd-070f-4d51-8017-c373abe6f339 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.824955] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12448ef-3ef5-4f08-b772-18810f63b8c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.827738] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275918, 'name': PowerOffVM_Task, 'duration_secs': 0.356837} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.839648] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.840030] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 965.840182] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275910, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.423863} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.840827] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-12b87ce2-ff38-455d-8019-f867fca5d321 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.843291] env[61824]: INFO nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f/OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f.vmdk to [datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e.vmdk. [ 965.843530] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Cleaning up location [datastore1] OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 965.843724] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_c0ba4739-b474-4c52-81bc-4dac2ec87e4f {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 965.845997] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1d072dfb-c5c3-4234-88b9-cc23d72231af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.852167] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a932dc1-478e-4fa0-8494-b90a51aca30f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.891273] env[61824]: INFO nova.compute.manager [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Detaching volume 279b78fe-a845-4106-9b4f-158ad3979315 [ 965.893540] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179645MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 965.893697] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.893885] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.896845] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 965.896845] env[61824]: value = "task-1275920" [ 965.896845] env[61824]: _type = "Task" [ 965.896845] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.908362] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037807} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.909258] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.909458] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.909799] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e.vmdk to [datastore1] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 965.910959] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-984f4c4a-31da-4bbb-bb2e-dcecb9528f27 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.919600] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 965.919600] env[61824]: value = "task-1275921" [ 965.919600] env[61824]: _type = "Task" [ 965.919600] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.931204] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275921, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.936634] env[61824]: INFO nova.virt.block_device [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Attempting to driver detach volume 279b78fe-a845-4106-9b4f-158ad3979315 from mountpoint /dev/sdb [ 965.936936] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 965.937261] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274181', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'name': 'volume-279b78fe-a845-4106-9b4f-158ad3979315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7be79422-c624-4b4a-884f-29b13b9d81d2', 'attached_at': '', 'detached_at': '', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'serial': '279b78fe-a845-4106-9b4f-158ad3979315'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 965.938246] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fc9a72-f65d-48b5-94b9-46a27bd238a8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.961251] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78dd144-d130-4e60-88a3-79562244af3b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.969079] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e64e69a-cc37-4037-825b-c7fb8a5ad802 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.990670] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4c1cfa4-34e5-46c6-8cce-43fa375cfbf7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.013301] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] The volume has not been displaced from its original location: [datastore1] volume-279b78fe-a845-4106-9b4f-158ad3979315/volume-279b78fe-a845-4106-9b4f-158ad3979315.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 966.018859] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 966.019246] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-823b229c-08d2-4757-9235-bb2790b86e0a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.041378] env[61824]: DEBUG oslo_vmware.api [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 966.041378] env[61824]: value = "task-1275922" [ 966.041378] env[61824]: _type = "Task" [ 966.041378] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.051222] env[61824]: DEBUG oslo_vmware.api [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275922, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.063394] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d8744b-dcaa-9d4d-3d22-bacb71a5e142, 'name': SearchDatastore_Task, 'duration_secs': 0.058675} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.063871] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.063990] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 966.064377] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.064588] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.064884] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.065392] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bcc155c8-d8d2-40c6-b392-918cca84d00d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.078838] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275917, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60945} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.083814] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 3256627d-35c8-4b75-b4a8-390eda0a150e/3256627d-35c8-4b75-b4a8-390eda0a150e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 966.084098] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 966.085600] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.085600] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 966.085600] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-390ed2fd-d7c0-449a-8421-759e10bf7b7a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.089867] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c03d9e0d-62c7-4baa-afdd-46707911e12b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.098057] env[61824]: DEBUG oslo_vmware.api [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275912, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.089239} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.099479] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.099816] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.100117] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.100366] env[61824]: INFO nova.compute.manager [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Took 3.25 seconds to destroy the instance on the hypervisor. [ 966.101047] env[61824]: DEBUG oslo.service.loopingcall [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.101470] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 966.101470] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52baf149-70e2-b45a-71d3-125c672e3a42" [ 966.101470] env[61824]: _type = "Task" [ 966.101470] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.103248] env[61824]: DEBUG nova.compute.manager [-] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 966.104355] env[61824]: DEBUG nova.network.neutron [-] [instance: e500f768-3ec8-48c7-a947-784e8665af15] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 966.105286] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 966.105286] env[61824]: value = "task-1275923" [ 966.105286] env[61824]: _type = "Task" [ 966.105286] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.121709] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52baf149-70e2-b45a-71d3-125c672e3a42, 'name': SearchDatastore_Task, 'duration_secs': 0.018557} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.125583] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.125862] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8656871-c4af-4f1c-9437-95e8f0dab12f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.135241] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 966.135241] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d882dd-4cff-6257-a15c-cb432c225a76" [ 966.135241] env[61824]: _type = "Task" [ 966.135241] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.145919] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d882dd-4cff-6257-a15c-cb432c225a76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.433850] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275921, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.464424] env[61824]: DEBUG nova.compute.manager [req-40681eee-74da-41d7-b857-95935d8655c2 req-318306f5-2b96-4802-9cf7-0b314df52824 service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-vif-deleted-6fe29327-b1fc-40ca-9f06-6bc80a449526 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.464537] env[61824]: INFO nova.compute.manager [req-40681eee-74da-41d7-b857-95935d8655c2 req-318306f5-2b96-4802-9cf7-0b314df52824 service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Neutron deleted interface 6fe29327-b1fc-40ca-9f06-6bc80a449526; detaching it from the instance and deleting it from the info cache [ 966.464820] env[61824]: DEBUG nova.network.neutron [req-40681eee-74da-41d7-b857-95935d8655c2 req-318306f5-2b96-4802-9cf7-0b314df52824 service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Updating instance_info_cache with network_info: [{"id": "10d4ea44-d470-4139-96d0-4c880d036717", "address": "fa:16:3e:65:a5:7f", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.146", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10d4ea44-d4", "ovs_interfaceid": "10d4ea44-d470-4139-96d0-4c880d036717", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "address": "fa:16:3e:6f:ef:7e", "network": {"id": "7af733a9-23dc-4dcd-add7-2188bb432a18", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-716900203", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.135", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2bf99f85-3a5c-47c6-a603-e215be6ab0bd", "external-id": "nsx-vlan-transportzone-855", "segmentation_id": 855, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2296c391-a8", "ovs_interfaceid": "2296c391-a8f5-4322-ac9f-7059d9ccb541", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.553895] env[61824]: DEBUG oslo_vmware.api [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275922, 'name': ReconfigVM_Task, 'duration_secs': 0.293004} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.554426] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 966.559161] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aac354d6-43a5-4851-99fd-498e4ed2c3ce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.578892] env[61824]: DEBUG oslo_vmware.api [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 966.578892] env[61824]: value = "task-1275924" [ 966.578892] env[61824]: _type = "Task" [ 966.578892] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.591144] env[61824]: DEBUG oslo_vmware.api [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275924, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.625294] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089279} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.625294] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 966.625294] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-790b4807-222a-443f-aabb-f11d83461cda {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.655832] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 3256627d-35c8-4b75-b4a8-390eda0a150e/3256627d-35c8-4b75-b4a8-390eda0a150e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 966.660785] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29fc9ab7-9ee4-4b4b-bad0-098045c6474d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.691280] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d882dd-4cff-6257-a15c-cb432c225a76, 'name': SearchDatastore_Task, 'duration_secs': 0.024207} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.693555] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.694207] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 82928174-1761-4296-bebe-7989c088d542/82928174-1761-4296-bebe-7989c088d542.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 966.694759] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 966.694759] env[61824]: value = "task-1275925" [ 966.694759] env[61824]: _type = "Task" [ 966.694759] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.695057] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de197a1a-38bd-45d8-9246-d8f035fd3a4e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.712229] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275925, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.714611] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 966.714611] env[61824]: value = "task-1275926" [ 966.714611] env[61824]: _type = "Task" [ 966.714611] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.728135] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275926, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.934217] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275921, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.935395] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.935715] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 7be79422-c624-4b4a-884f-29b13b9d81d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.935827] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance c5da84f7-4ee7-4537-a3f5-b95ee513c4dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.935914] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance e500f768-3ec8-48c7-a947-784e8665af15 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.936086] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 21759f11-80e9-419d-9f65-ca4767d1593b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.936246] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 44ec5c22-d080-45d1-aa21-080af4c317c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.936394] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 3256627d-35c8-4b75-b4a8-390eda0a150e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.936537] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 82928174-1761-4296-bebe-7989c088d542 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 966.936867] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 966.937063] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 966.951843] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 966.952150] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 966.952343] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleting the datastore file [datastore1] 44ec5c22-d080-45d1-aa21-080af4c317c5 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.956029] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23f7500f-ac68-458c-806a-9211bdb76921 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.961306] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 966.961306] env[61824]: value = "task-1275927" [ 966.961306] env[61824]: _type = "Task" [ 966.961306] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.969855] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36bc84c3-a756-4b89-910a-b2980f5b628f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.978834] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.989041] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a6948d-1fce-49fd-86df-bce51d6e093c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.026339] env[61824]: DEBUG nova.compute.manager [req-40681eee-74da-41d7-b857-95935d8655c2 req-318306f5-2b96-4802-9cf7-0b314df52824 service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Detach interface failed, port_id=6fe29327-b1fc-40ca-9f06-6bc80a449526, reason: Instance e500f768-3ec8-48c7-a947-784e8665af15 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 967.092311] env[61824]: DEBUG oslo_vmware.api [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275924, 'name': ReconfigVM_Task, 'duration_secs': 0.173609} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.092482] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274181', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'name': 'volume-279b78fe-a845-4106-9b4f-158ad3979315', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '7be79422-c624-4b4a-884f-29b13b9d81d2', 'attached_at': '', 'detached_at': '', 'volume_id': '279b78fe-a845-4106-9b4f-158ad3979315', 'serial': '279b78fe-a845-4106-9b4f-158ad3979315'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 967.097393] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ead68b-8092-46ab-bc68-890e0c2b7456 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.108153] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b945090e-4c4a-4dcd-984f-508b9ce9f875 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.146843] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035a1de0-3e81-4a26-abe1-adc4c96edb49 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.157940] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c301eb70-1e42-4f73-9613-4a582da556c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.176043] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.217028] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275925, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.233356] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275926, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.397142] env[61824]: DEBUG nova.network.neutron [-] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.450802] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275921, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.477659] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.653331] env[61824]: DEBUG nova.objects.instance [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'flavor' on Instance uuid 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.680043] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.709505] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275925, 'name': ReconfigVM_Task, 'duration_secs': 0.811729} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.709914] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 3256627d-35c8-4b75-b4a8-390eda0a150e/3256627d-35c8-4b75-b4a8-390eda0a150e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 967.710689] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-459f49f4-e1b9-4979-a156-1db61ecae241 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.721198] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 967.721198] env[61824]: value = "task-1275929" [ 967.721198] env[61824]: _type = "Task" [ 967.721198] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.730025] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275926, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.741217} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.730025] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 82928174-1761-4296-bebe-7989c088d542/82928174-1761-4296-bebe-7989c088d542.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 967.730284] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 967.730326] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ddf7596-8d8d-4404-b3ef-eeb4f4e795d8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.735646] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275929, 'name': Rename_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.739995] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 967.739995] env[61824]: value = "task-1275930" [ 967.739995] env[61824]: _type = "Task" [ 967.739995] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.747885] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275930, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.900732] env[61824]: INFO nova.compute.manager [-] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Took 1.80 seconds to deallocate network for instance. [ 967.932924] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275921, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.976961] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275927, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.184327] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 968.184513] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.291s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.232613] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275929, 'name': Rename_Task, 'duration_secs': 0.194199} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.232894] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 968.233170] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04c6f53b-e3d4-4a16-8f46-6696223dfaf2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.240221] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 968.240221] env[61824]: value = "task-1275931" [ 968.240221] env[61824]: _type = "Task" [ 968.240221] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.251631] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.255982] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275930, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073292} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.255982] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.255982] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e93d5940-3454-485a-a008-205e41256853 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.280042] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 82928174-1761-4296-bebe-7989c088d542/82928174-1761-4296-bebe-7989c088d542.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.283144] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d493fe8-4974-44d1-a160-ba6b57523a83 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.301349] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 968.301349] env[61824]: value = "task-1275932" [ 968.301349] env[61824]: _type = "Task" [ 968.301349] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.310255] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275932, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.411722] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.412055] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.412343] env[61824]: DEBUG nova.objects.instance [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lazy-loading 'resources' on Instance uuid e500f768-3ec8-48c7-a947-784e8665af15 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.432452] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275921, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.33811} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.432589] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e/0da722b8-2f9f-4a79-ada2-0ce4f4dd247e.vmdk to [datastore1] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 968.433412] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cac2c84-7793-4875-90c7-59033c6935d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.455559] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.456635] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2fd0a50-555e-41fa-9f33-7e69485f8985 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.480575] env[61824]: DEBUG oslo_vmware.api [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275927, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.358087} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.481996] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.482243] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 968.482461] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 968.482674] env[61824]: INFO nova.compute.manager [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Took 3.21 seconds to destroy the instance on the hypervisor. [ 968.482961] env[61824]: DEBUG oslo.service.loopingcall [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.483286] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 968.483286] env[61824]: value = "task-1275933" [ 968.483286] env[61824]: _type = "Task" [ 968.483286] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.483507] env[61824]: DEBUG nova.compute.manager [-] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 968.483619] env[61824]: DEBUG nova.network.neutron [-] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 968.494803] env[61824]: DEBUG nova.compute.manager [req-b58a2e05-5acb-446f-a2a2-0e1ff1b0b616 req-dbef28c5-544a-4eb3-aed0-4b91e320fd3a service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-vif-deleted-2296c391-a8f5-4322-ac9f-7059d9ccb541 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.495017] env[61824]: DEBUG nova.compute.manager [req-b58a2e05-5acb-446f-a2a2-0e1ff1b0b616 req-dbef28c5-544a-4eb3-aed0-4b91e320fd3a service nova] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Received event network-vif-deleted-10d4ea44-d470-4139-96d0-4c880d036717 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.498657] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275933, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.661810] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a65d4d15-eeb5-449e-b3ea-e4bfbd26cd48 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.283s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.753390] env[61824]: DEBUG oslo_vmware.api [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275931, 'name': PowerOnVM_Task, 'duration_secs': 0.484931} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.753669] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 968.753876] env[61824]: INFO nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Took 9.48 seconds to spawn the instance on the hypervisor. [ 968.754063] env[61824]: DEBUG nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 968.755106] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4f2afc-fdaa-4652-a7ba-083d537a3dc2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.812491] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275932, 'name': ReconfigVM_Task, 'duration_secs': 0.485291} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.813134] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 82928174-1761-4296-bebe-7989c088d542/82928174-1761-4296-bebe-7989c088d542.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.814170] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5a84fb2a-f487-44c2-a377-aab119252331 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.820847] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 968.820847] env[61824]: value = "task-1275934" [ 968.820847] env[61824]: _type = "Task" [ 968.820847] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.830897] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275934, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.995435] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275933, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.049176] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee38fcc-813e-43af-a522-ba1e884351e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.056935] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeccdff5-e497-4e52-a74a-b7f78bd6d8f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.098412] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5093d4-6243-4535-b1f2-06535d686cb2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.106786] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c589b4-a2e2-4b5b-a7fd-783ecd5f013d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.120206] env[61824]: DEBUG nova.compute.provider_tree [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.280474] env[61824]: INFO nova.compute.manager [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Took 21.68 seconds to build instance. [ 969.313836] env[61824]: DEBUG nova.network.neutron [-] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.332679] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275934, 'name': Rename_Task, 'duration_secs': 0.344977} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.333186] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.333447] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cd7f8a1-7d10-4147-a7fa-098b82be87df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.340884] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 969.340884] env[61824]: value = "task-1275935" [ 969.340884] env[61824]: _type = "Task" [ 969.340884] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.351475] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275935, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.485475] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.485475] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.497032] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275933, 'name': ReconfigVM_Task, 'duration_secs': 0.546023} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.498037] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 21759f11-80e9-419d-9f65-ca4767d1593b/21759f11-80e9-419d-9f65-ca4767d1593b.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.498684] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afe3d9a0-44d7-4689-95fb-6744e54cb5f4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.506162] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 969.506162] env[61824]: value = "task-1275936" [ 969.506162] env[61824]: _type = "Task" [ 969.506162] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.515642] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275936, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.623658] env[61824]: DEBUG nova.scheduler.client.report [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.788498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-901759a6-e44d-4135-889f-6184974f333a tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.197s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.818435] env[61824]: INFO nova.compute.manager [-] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Took 1.33 seconds to deallocate network for instance. [ 969.851395] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275935, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.878152] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.878152] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.878152] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.878498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.878498] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.881169] env[61824]: INFO nova.compute.manager [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Terminating instance [ 969.883327] env[61824]: DEBUG nova.compute.manager [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.883564] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.884662] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6482794f-dc8b-48ae-a906-222522f4ba6e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.894291] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.894653] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca5a17d7-bbaa-46be-b195-ec46d477ee43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.901414] env[61824]: DEBUG oslo_vmware.api [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 969.901414] env[61824]: value = "task-1275937" [ 969.901414] env[61824]: _type = "Task" [ 969.901414] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.911086] env[61824]: DEBUG oslo_vmware.api [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.950665] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "3256627d-35c8-4b75-b4a8-390eda0a150e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.950941] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.951294] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "3256627d-35c8-4b75-b4a8-390eda0a150e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.951576] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.951823] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.955384] env[61824]: INFO nova.compute.manager [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Terminating instance [ 969.958517] env[61824]: DEBUG nova.compute.manager [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.958890] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 969.960378] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35543c6b-d901-4c80-81f4-e22703369e60 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.970949] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.971422] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c0395ab-08e5-4523-9ebd-ecca0898abc9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.979070] env[61824]: DEBUG oslo_vmware.api [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 969.979070] env[61824]: value = "task-1275938" [ 969.979070] env[61824]: _type = "Task" [ 969.979070] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.990995] env[61824]: DEBUG nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 969.995418] env[61824]: DEBUG oslo_vmware.api [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.018612] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275936, 'name': Rename_Task, 'duration_secs': 0.386393} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.018955] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 970.019260] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20882477-de90-4a85-9b2b-3004f89a4699 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.026931] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 970.026931] env[61824]: value = "task-1275939" [ 970.026931] env[61824]: _type = "Task" [ 970.026931] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.036210] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.129726] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.717s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.154465] env[61824]: INFO nova.scheduler.client.report [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Deleted allocations for instance e500f768-3ec8-48c7-a947-784e8665af15 [ 970.325565] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.325872] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.326149] env[61824]: DEBUG nova.objects.instance [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'resources' on Instance uuid 44ec5c22-d080-45d1-aa21-080af4c317c5 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.352075] env[61824]: DEBUG oslo_vmware.api [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275935, 'name': PowerOnVM_Task, 'duration_secs': 0.628457} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.352075] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.352397] env[61824]: INFO nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Took 8.75 seconds to spawn the instance on the hypervisor. [ 970.352397] env[61824]: DEBUG nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.353160] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb163aad-c3f4-4504-a9eb-640a2ac3c697 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.411642] env[61824]: DEBUG oslo_vmware.api [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275937, 'name': PowerOffVM_Task, 'duration_secs': 0.295221} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.411957] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.412156] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.412432] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d72c0b7f-eb58-4a84-9336-b69afccecd43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.487038] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.487413] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.487541] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleting the datastore file [datastore1] 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.490843] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7780bbc5-fb8f-4fce-8026-0839a1ed7a38 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.492762] env[61824]: DEBUG oslo_vmware.api [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275938, 'name': PowerOffVM_Task, 'duration_secs': 0.292271} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.492998] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.493196] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.493819] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-114f010e-158f-4e3f-b25c-f88c9a1d9a59 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.499370] env[61824]: DEBUG oslo_vmware.api [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 970.499370] env[61824]: value = "task-1275941" [ 970.499370] env[61824]: _type = "Task" [ 970.499370] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.510454] env[61824]: DEBUG oslo_vmware.api [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275941, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.513638] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.539627] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275939, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.576073] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.576073] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.576263] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleting the datastore file [datastore2] 3256627d-35c8-4b75-b4a8-390eda0a150e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.576585] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6e26b28-d19b-4c6e-85e4-d7cf4cc75cc0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.588566] env[61824]: DEBUG oslo_vmware.api [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 970.588566] env[61824]: value = "task-1275943" [ 970.588566] env[61824]: _type = "Task" [ 970.588566] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.603595] env[61824]: DEBUG oslo_vmware.api [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275943, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.663028] env[61824]: DEBUG oslo_concurrency.lockutils [None req-502e6884-70e5-4ef1-a476-810dcac3d9f8 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "e500f768-3ec8-48c7-a947-784e8665af15" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.816s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.871600] env[61824]: INFO nova.compute.manager [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Took 14.25 seconds to build instance. [ 970.949456] env[61824]: DEBUG nova.compute.manager [req-d4c10be3-441e-44af-a59c-0ca20c3f91a5 req-88252700-aebd-4012-a8a3-b8b2ccd20dec service nova] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Received event network-vif-deleted-7832e699-039f-4690-9e62-cdb49ebee107 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.989574] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c31070-cce6-49df-9492-1fbce581c107 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.997920] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73bdb9c1-51ea-4e4f-98aa-0478d6f8f782 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.012727] env[61824]: DEBUG oslo_vmware.api [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275941, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193076} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.037437] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.037437] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.037625] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.037900] env[61824]: INFO nova.compute.manager [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Took 1.15 seconds to destroy the instance on the hypervisor. [ 971.038104] env[61824]: DEBUG oslo.service.loopingcall [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.039081] env[61824]: DEBUG nova.compute.manager [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.039190] env[61824]: DEBUG nova.network.neutron [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 971.044560] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b53af9-94c0-4435-86b9-9dc0ad3debad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.053586] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275939, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.056879] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1b2ff6-7b8a-4287-b3c3-fd2de4e3e6bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.071782] env[61824]: DEBUG nova.compute.provider_tree [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.100757] env[61824]: DEBUG oslo_vmware.api [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275943, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169601} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.101053] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.101599] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.101599] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.101599] env[61824]: INFO nova.compute.manager [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 971.101862] env[61824]: DEBUG oslo.service.loopingcall [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.102386] env[61824]: DEBUG nova.compute.manager [-] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.102479] env[61824]: DEBUG nova.network.neutron [-] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 971.375574] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f4a33e56-b518-4633-b120-fdbeaa3ad3a9 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "82928174-1761-4296-bebe-7989c088d542" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.765s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.548437] env[61824]: DEBUG oslo_vmware.api [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275939, 'name': PowerOnVM_Task, 'duration_secs': 1.246746} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.548769] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.576033] env[61824]: DEBUG nova.scheduler.client.report [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.691901] env[61824]: DEBUG nova.compute.manager [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.693216] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7c9358-1fc0-4b71-9144-b70a8e43720f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.978640] env[61824]: DEBUG nova.network.neutron [-] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.085025] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.757s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.085025] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.571s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.087407] env[61824]: INFO nova.compute.claims [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 972.159029] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a94c328d-c649-4941-a959-89d8ddb918fe tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "interface-82928174-1761-4296-bebe-7989c088d542-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.159442] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a94c328d-c649-4941-a959-89d8ddb918fe tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "interface-82928174-1761-4296-bebe-7989c088d542-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.159733] env[61824]: DEBUG nova.objects.instance [None req-a94c328d-c649-4941-a959-89d8ddb918fe tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lazy-loading 'flavor' on Instance uuid 82928174-1761-4296-bebe-7989c088d542 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.195911] env[61824]: INFO nova.scheduler.client.report [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted allocations for instance 44ec5c22-d080-45d1-aa21-080af4c317c5 [ 972.210148] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b6a7129-f6d5-4867-9563-dd83d8d57bf2 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.705s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.214650] env[61824]: DEBUG nova.network.neutron [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.482085] env[61824]: INFO nova.compute.manager [-] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Took 1.38 seconds to deallocate network for instance. [ 972.664583] env[61824]: DEBUG nova.objects.instance [None req-a94c328d-c649-4941-a959-89d8ddb918fe tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lazy-loading 'pci_requests' on Instance uuid 82928174-1761-4296-bebe-7989c088d542 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.706103] env[61824]: DEBUG oslo_concurrency.lockutils [None req-7bfc2ab9-f87d-4093-8655-f45fceb82bd0 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "44ec5c22-d080-45d1-aa21-080af4c317c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.442s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.716050] env[61824]: INFO nova.compute.manager [-] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Took 1.68 seconds to deallocate network for instance. [ 972.990963] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.167107] env[61824]: DEBUG nova.objects.base [None req-a94c328d-c649-4941-a959-89d8ddb918fe tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Object Instance<82928174-1761-4296-bebe-7989c088d542> lazy-loaded attributes: flavor,pci_requests {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 973.167256] env[61824]: DEBUG nova.network.neutron [None req-a94c328d-c649-4941-a959-89d8ddb918fe tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 973.196100] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-017af940-ec27-4110-8985-b56c461c463a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.205532] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da2ae10-70ff-4085-9fe6-171bca44d998 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.238652] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.244786] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b320571b-071e-4c06-b364-cd038e0c63f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.250234] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169f0524-fad8-48b8-8478-b233bbb89266 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.263679] env[61824]: DEBUG nova.compute.provider_tree [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.281885] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a94c328d-c649-4941-a959-89d8ddb918fe tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "interface-82928174-1761-4296-bebe-7989c088d542-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.122s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.394485] env[61824]: DEBUG nova.compute.manager [req-65b9d43b-2f8a-4332-8fcf-a8635c3faacb req-0cad689d-d3e4-41b1-ba83-cc813bdeddb3 service nova] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Received event network-vif-deleted-22e2de70-07bb-4d94-8624-d6082438907b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.394728] env[61824]: DEBUG nova.compute.manager [req-65b9d43b-2f8a-4332-8fcf-a8635c3faacb req-0cad689d-d3e4-41b1-ba83-cc813bdeddb3 service nova] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Received event network-vif-deleted-a66cb8c8-5826-4e02-ad96-b2bbea85d23b {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.767303] env[61824]: DEBUG nova.scheduler.client.report [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.898059] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.898356] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.272392] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.187s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.272935] env[61824]: DEBUG nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.275486] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.285s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.275697] env[61824]: DEBUG nova.objects.instance [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'resources' on Instance uuid 3256627d-35c8-4b75-b4a8-390eda0a150e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.401168] env[61824]: DEBUG nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 974.528415] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.528650] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.640398] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762b481d-6cce-4b5d-9ea2-9d702eb70160 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.647150] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a04b964-ab7a-43ce-8827-4d8369d7d593 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Suspending the VM {{(pid=61824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 974.647580] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-dd7d74f4-1c22-4f5c-864c-09d59162a677 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.654020] env[61824]: DEBUG oslo_vmware.api [None req-4a04b964-ab7a-43ce-8827-4d8369d7d593 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 974.654020] env[61824]: value = "task-1275944" [ 974.654020] env[61824]: _type = "Task" [ 974.654020] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.661924] env[61824]: DEBUG oslo_vmware.api [None req-4a04b964-ab7a-43ce-8827-4d8369d7d593 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275944, 'name': SuspendVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.779269] env[61824]: DEBUG nova.compute.utils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.782783] env[61824]: DEBUG nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.783953] env[61824]: DEBUG nova.network.neutron [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 974.851165] env[61824]: DEBUG nova.policy [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad4e8224518048e583d27b94721b3c22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd21256931aa54010b72beca4fb798f63', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 974.918270] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace0f7e6-94cb-47fa-9c8f-b502ca94c353 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.928302] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61ffac1-2053-4bbb-a4c2-4051f1b65d1c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.936022] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.965531] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dfc7c6-120c-4077-bdcd-f5361ef85079 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.973461] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396716f3-7201-4f70-8784-9b0fc7bda61c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.988138] env[61824]: DEBUG nova.compute.provider_tree [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.030986] env[61824]: DEBUG nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 975.165662] env[61824]: DEBUG oslo_vmware.api [None req-4a04b964-ab7a-43ce-8827-4d8369d7d593 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275944, 'name': SuspendVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.183451] env[61824]: DEBUG nova.network.neutron [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Successfully created port: 87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.228571] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "82928174-1761-4296-bebe-7989c088d542" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.228963] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "82928174-1761-4296-bebe-7989c088d542" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.229124] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "82928174-1761-4296-bebe-7989c088d542-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.229343] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "82928174-1761-4296-bebe-7989c088d542-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.229536] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "82928174-1761-4296-bebe-7989c088d542-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.232533] env[61824]: INFO nova.compute.manager [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Terminating instance [ 975.234260] env[61824]: DEBUG nova.compute.manager [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.234530] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 975.235353] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83e9f3d-663b-4853-80f3-afa97fcf33bf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.248845] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 975.249146] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-218a25fe-ea8f-43aa-84ae-fbd9420f903e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.260811] env[61824]: DEBUG oslo_vmware.api [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 975.260811] env[61824]: value = "task-1275945" [ 975.260811] env[61824]: _type = "Task" [ 975.260811] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.272448] env[61824]: DEBUG oslo_vmware.api [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.286286] env[61824]: DEBUG nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.490594] env[61824]: DEBUG nova.scheduler.client.report [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 975.549432] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.663924] env[61824]: DEBUG oslo_vmware.api [None req-4a04b964-ab7a-43ce-8827-4d8369d7d593 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275944, 'name': SuspendVM_Task, 'duration_secs': 0.665896} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.664210] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a04b964-ab7a-43ce-8827-4d8369d7d593 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Suspended the VM {{(pid=61824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 975.664394] env[61824]: DEBUG nova.compute.manager [None req-4a04b964-ab7a-43ce-8827-4d8369d7d593 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.665128] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5aa7ede-a356-4b0b-8977-f345550697a5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.770988] env[61824]: DEBUG oslo_vmware.api [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275945, 'name': PowerOffVM_Task, 'duration_secs': 0.178127} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.771202] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 975.771378] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 975.771626] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7e84258-d66c-4bd9-970b-ad81f0dcc225 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.857965] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 975.858355] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 975.858412] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Deleting the datastore file [datastore2] 82928174-1761-4296-bebe-7989c088d542 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.858627] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de599770-b875-46ed-850e-96733c323a48 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.864972] env[61824]: DEBUG oslo_vmware.api [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for the task: (returnval){ [ 975.864972] env[61824]: value = "task-1275947" [ 975.864972] env[61824]: _type = "Task" [ 975.864972] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.872411] env[61824]: DEBUG oslo_vmware.api [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.996021] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.720s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.997870] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.759s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.998125] env[61824]: DEBUG nova.objects.instance [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'resources' on Instance uuid 7be79422-c624-4b4a-884f-29b13b9d81d2 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.018092] env[61824]: INFO nova.scheduler.client.report [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted allocations for instance 3256627d-35c8-4b75-b4a8-390eda0a150e [ 976.296223] env[61824]: DEBUG nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.320651] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.320903] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.321080] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.321272] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.321419] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.321568] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.321780] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.321940] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.322126] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.322295] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.322468] env[61824]: DEBUG nova.virt.hardware [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.323688] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af23a7c-2142-4110-aabc-7957451d56e7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.331509] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d78a44ba-b142-4bf0-9e74-6428be3d01b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.373755] env[61824]: DEBUG oslo_vmware.api [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Task: {'id': task-1275947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148284} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.373988] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.374193] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.374370] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.374545] env[61824]: INFO nova.compute.manager [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] [instance: 82928174-1761-4296-bebe-7989c088d542] Took 1.14 seconds to destroy the instance on the hypervisor. [ 976.374780] env[61824]: DEBUG oslo.service.loopingcall [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.374968] env[61824]: DEBUG nova.compute.manager [-] [instance: 82928174-1761-4296-bebe-7989c088d542] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.375113] env[61824]: DEBUG nova.network.neutron [-] [instance: 82928174-1761-4296-bebe-7989c088d542] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.527038] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f5e04f9-7b17-4a6b-b55a-0b7d2c049078 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "3256627d-35c8-4b75-b4a8-390eda0a150e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.576s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.616813] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1fbba5-de83-4e58-b415-775a57a70561 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.629248] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d18d534-ffd2-4457-8926-75856fe0659c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.664913] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d859da6-c68d-4f34-8200-e383d3804d19 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.672636] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a9116fc-14c7-4eb2-9a60-18a37005269f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.690740] env[61824]: DEBUG nova.compute.provider_tree [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.764694] env[61824]: DEBUG nova.compute.manager [req-819a3b0d-5977-417b-9d42-3510e6dc8dc2 req-edc47a47-486a-4971-91ea-426950e45936 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Received event network-vif-deleted-44ea9f2e-3382-431f-b83a-590af6c7f511 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.764903] env[61824]: INFO nova.compute.manager [req-819a3b0d-5977-417b-9d42-3510e6dc8dc2 req-edc47a47-486a-4971-91ea-426950e45936 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Neutron deleted interface 44ea9f2e-3382-431f-b83a-590af6c7f511; detaching it from the instance and deleting it from the info cache [ 976.765099] env[61824]: DEBUG nova.network.neutron [req-819a3b0d-5977-417b-9d42-3510e6dc8dc2 req-edc47a47-486a-4971-91ea-426950e45936 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.960044] env[61824]: DEBUG nova.network.neutron [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Successfully updated port: 87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.192036] env[61824]: INFO nova.compute.manager [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Resuming [ 977.192036] env[61824]: DEBUG nova.objects.instance [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lazy-loading 'flavor' on Instance uuid 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.194249] env[61824]: DEBUG nova.scheduler.client.report [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 977.236424] env[61824]: DEBUG nova.network.neutron [-] [instance: 82928174-1761-4296-bebe-7989c088d542] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.267870] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5e93390-b9a0-4945-9fe8-9063eb099b0c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.278261] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2a0703-1288-4cad-a59e-a615f92a93c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.302540] env[61824]: DEBUG nova.compute.manager [req-819a3b0d-5977-417b-9d42-3510e6dc8dc2 req-edc47a47-486a-4971-91ea-426950e45936 service nova] [instance: 82928174-1761-4296-bebe-7989c088d542] Detach interface failed, port_id=44ea9f2e-3382-431f-b83a-590af6c7f511, reason: Instance 82928174-1761-4296-bebe-7989c088d542 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 977.416074] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.416364] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.416583] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.416771] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.416971] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.419037] env[61824]: INFO nova.compute.manager [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Terminating instance [ 977.420720] env[61824]: DEBUG nova.compute.manager [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 977.420911] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 977.421758] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da277bf-04e6-44b3-9fde-6b5dcc3acddc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.431196] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 977.431434] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc2adc78-ea4b-49e0-b6e8-3941b54f1801 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.448710] env[61824]: DEBUG oslo_vmware.api [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 977.448710] env[61824]: value = "task-1275948" [ 977.448710] env[61824]: _type = "Task" [ 977.448710] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.456028] env[61824]: DEBUG oslo_vmware.api [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275948, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.463634] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.463773] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.463932] env[61824]: DEBUG nova.network.neutron [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 977.700221] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.705026] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.768s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.705629] env[61824]: INFO nova.compute.claims [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.722518] env[61824]: INFO nova.scheduler.client.report [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted allocations for instance 7be79422-c624-4b4a-884f-29b13b9d81d2 [ 977.739037] env[61824]: INFO nova.compute.manager [-] [instance: 82928174-1761-4296-bebe-7989c088d542] Took 1.36 seconds to deallocate network for instance. [ 977.959229] env[61824]: DEBUG oslo_vmware.api [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275948, 'name': PowerOffVM_Task, 'duration_secs': 0.200127} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.959553] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 977.959731] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 977.960019] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8603fa8c-e0d0-4ba0-90c8-ce8d8e5f06b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.019267] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.019517] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.019684] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleting the datastore file [datastore2] c5da84f7-4ee7-4537-a3f5-b95ee513c4dd {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.020548] env[61824]: DEBUG nova.network.neutron [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 978.022396] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-550e19b7-3fed-4c3e-9647-c467af0da067 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.028821] env[61824]: DEBUG oslo_vmware.api [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 978.028821] env[61824]: value = "task-1275950" [ 978.028821] env[61824]: _type = "Task" [ 978.028821] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.036474] env[61824]: DEBUG oslo_vmware.api [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275950, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.209892] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.210078] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquired lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.210274] env[61824]: DEBUG nova.network.neutron [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 978.225936] env[61824]: DEBUG nova.network.neutron [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [{"id": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "address": "fa:16:3e:83:f8:c1", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ec4b1c-0d", "ovs_interfaceid": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.229392] env[61824]: DEBUG oslo_concurrency.lockutils [None req-190e74f6-ebf6-46c1-89db-df93c45a1db5 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "7be79422-c624-4b4a-884f-29b13b9d81d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.352s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.245147] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.538216] env[61824]: DEBUG oslo_vmware.api [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275950, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131192} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.538474] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 978.538662] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 978.538841] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 978.539070] env[61824]: INFO nova.compute.manager [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Took 1.12 seconds to destroy the instance on the hypervisor. [ 978.539326] env[61824]: DEBUG oslo.service.loopingcall [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.539523] env[61824]: DEBUG nova.compute.manager [-] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 978.539619] env[61824]: DEBUG nova.network.neutron [-] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 978.731268] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.731552] env[61824]: DEBUG nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Instance network_info: |[{"id": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "address": "fa:16:3e:83:f8:c1", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ec4b1c-0d", "ovs_interfaceid": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.732130] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:f8:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87ec4b1c-0d02-46ba-878d-b392ae74f472', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.743461] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating folder: Project (d21256931aa54010b72beca4fb798f63). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 978.746747] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d0f619e-bb17-4922-ac5e-b3888b05b960 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.756933] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created folder: Project (d21256931aa54010b72beca4fb798f63) in parent group-v274074. [ 978.757030] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating folder: Instances. Parent ref: group-v274214. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 978.757278] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f46c9a49-6d86-4cf7-911e-9ebac3f78390 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.766108] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created folder: Instances in parent group-v274214. [ 978.766359] env[61824]: DEBUG oslo.service.loopingcall [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.768549] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.768779] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-237cb45f-b9ba-4b1f-ad26-e82959851505 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.790323] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.790323] env[61824]: value = "task-1275953" [ 978.790323] env[61824]: _type = "Task" [ 978.790323] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.796610] env[61824]: DEBUG nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Received event network-vif-plugged-87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.797200] env[61824]: DEBUG oslo_concurrency.lockutils [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] Acquiring lock "8aa00861-bca6-4861-a5e5-0c538155c9ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.797423] env[61824]: DEBUG oslo_concurrency.lockutils [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.797626] env[61824]: DEBUG oslo_concurrency.lockutils [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.797668] env[61824]: DEBUG nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] No waiting events found dispatching network-vif-plugged-87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 978.797880] env[61824]: WARNING nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Received unexpected event network-vif-plugged-87ec4b1c-0d02-46ba-878d-b392ae74f472 for instance with vm_state building and task_state spawning. [ 978.798066] env[61824]: DEBUG nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Received event network-changed-87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.798187] env[61824]: DEBUG nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Refreshing instance network info cache due to event network-changed-87ec4b1c-0d02-46ba-878d-b392ae74f472. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 978.798359] env[61824]: DEBUG oslo_concurrency.lockutils [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] Acquiring lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.798973] env[61824]: DEBUG oslo_concurrency.lockutils [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] Acquired lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.798973] env[61824]: DEBUG nova.network.neutron [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Refreshing network info cache for port 87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 978.805170] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275953, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.852782] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc5cc89-2b18-4d29-b88d-55d6c4685d93 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.863997] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bc14cc-9c31-4d64-9b0b-1e971b0fc745 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.897078] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92733d66-6d0e-4302-b759-d1e0276317b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.904578] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1685160f-75bf-47f8-a780-ed49e1df4ea3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.917703] env[61824]: DEBUG nova.compute.provider_tree [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.976176] env[61824]: DEBUG nova.network.neutron [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [{"id": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "address": "fa:16:3e:8e:8d:c2", "network": {"id": "bfc00206-1338-471e-860c-a61d2bc76b87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-46500319-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc352363e7f94dc6b7a99db8aca565a6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "55764410-260e-4339-a020-6b30995584bf", "external-id": "nsx-vlan-transportzone-551", "segmentation_id": 551, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45dc4201-9f", "ovs_interfaceid": "45dc4201-9fb0-4060-ba24-f376e69a6e45", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.269800] env[61824]: DEBUG nova.network.neutron [-] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.301650] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275953, 'name': CreateVM_Task, 'duration_secs': 0.416809} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.302670] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.303735] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.303735] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.303842] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.304234] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24fbe4d6-d5e1-4f0e-bb4d-8b452dd9d4cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.310707] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 979.310707] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521a3c4b-bbbd-d148-c03b-431875868c26" [ 979.310707] env[61824]: _type = "Task" [ 979.310707] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.318974] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521a3c4b-bbbd-d148-c03b-431875868c26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.420472] env[61824]: DEBUG nova.scheduler.client.report [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.479409] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Releasing lock "refresh_cache-21759f11-80e9-419d-9f65-ca4767d1593b" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.480462] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cacb6e-107e-4762-92c1-353fa070f6f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.487229] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Resuming the VM {{(pid=61824) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 979.487299] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4a0e6de-f476-4d47-b546-bc29b402d952 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.495728] env[61824]: DEBUG oslo_vmware.api [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 979.495728] env[61824]: value = "task-1275954" [ 979.495728] env[61824]: _type = "Task" [ 979.495728] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.506902] env[61824]: DEBUG oslo_vmware.api [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.522695] env[61824]: DEBUG nova.network.neutron [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updated VIF entry in instance network info cache for port 87ec4b1c-0d02-46ba-878d-b392ae74f472. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 979.523400] env[61824]: DEBUG nova.network.neutron [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [{"id": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "address": "fa:16:3e:83:f8:c1", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ec4b1c-0d", "ovs_interfaceid": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.774815] env[61824]: INFO nova.compute.manager [-] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Took 1.24 seconds to deallocate network for instance. [ 979.822909] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521a3c4b-bbbd-d148-c03b-431875868c26, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.823200] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.823585] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.823860] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.824036] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.824230] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.824510] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41e89cc3-ce56-4623-9b0c-7c4fc6fd8ec5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.834926] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.835168] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.836547] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bad074e9-3104-4400-bd15-e86b000fa303 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.843152] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 979.843152] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e77217-8fdc-d5e8-9078-48be916c029b" [ 979.843152] env[61824]: _type = "Task" [ 979.843152] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.852109] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e77217-8fdc-d5e8-9078-48be916c029b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.926535] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.223s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.927145] env[61824]: DEBUG nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 979.929812] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.380s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.931321] env[61824]: INFO nova.compute.claims [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.005857] env[61824]: DEBUG oslo_vmware.api [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275954, 'name': PowerOnVM_Task, 'duration_secs': 0.480579} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.006423] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Resumed the VM {{(pid=61824) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 980.006805] env[61824]: DEBUG nova.compute.manager [None req-d5272b7d-c29b-466e-8d63-d9b144620694 tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 980.007509] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f31776b-e9e8-4120-8263-fa90aef1eef9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.026116] env[61824]: DEBUG oslo_concurrency.lockutils [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] Releasing lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.026430] env[61824]: DEBUG nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Received event network-vif-deleted-a50e03ba-154c-4623-9e07-df8b06e3233e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.026616] env[61824]: INFO nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Neutron deleted interface a50e03ba-154c-4623-9e07-df8b06e3233e; detaching it from the instance and deleting it from the info cache [ 980.026796] env[61824]: DEBUG nova.network.neutron [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.281891] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.354560] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e77217-8fdc-d5e8-9078-48be916c029b, 'name': SearchDatastore_Task, 'duration_secs': 0.011064} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.355390] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21e28f43-d98f-4db7-852e-06ba2fc12ec3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.360801] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 980.360801] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c1e69d-72ee-ea39-5d00-769d48f79ae3" [ 980.360801] env[61824]: _type = "Task" [ 980.360801] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.368443] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c1e69d-72ee-ea39-5d00-769d48f79ae3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.435864] env[61824]: DEBUG nova.compute.utils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.439407] env[61824]: DEBUG nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 980.439563] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 980.496537] env[61824]: DEBUG nova.policy [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8835a9932a0c42d68e6640bcc3b4978c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '773a259f132a42479464e120df314bb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.529535] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-17eabdab-c136-45c1-91b0-59b8d9f0d77a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.539135] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2edd5b-ef65-4787-8ef8-e6c7f31f2aad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.565863] env[61824]: DEBUG nova.compute.manager [req-29cd17c7-c847-4273-85ed-3b218a698407 req-f2d9aa91-89ef-4a90-980a-6ef35f6adf4d service nova] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Detach interface failed, port_id=a50e03ba-154c-4623-9e07-df8b06e3233e, reason: Instance c5da84f7-4ee7-4537-a3f5-b95ee513c4dd could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.679376] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.679376] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.877315] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c1e69d-72ee-ea39-5d00-769d48f79ae3, 'name': SearchDatastore_Task, 'duration_secs': 0.063236} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.877315] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.877315] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 8aa00861-bca6-4861-a5e5-0c538155c9ab/8aa00861-bca6-4861-a5e5-0c538155c9ab.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.877315] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2578f5a-777c-474d-88f8-4198cf18f8fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.880988] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Successfully created port: 841a19df-b0cc-4ca3-a346-b9f56cbd1883 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.889755] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 980.889755] env[61824]: value = "task-1275955" [ 980.889755] env[61824]: _type = "Task" [ 980.889755] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.902176] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275955, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.943227] env[61824]: DEBUG nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.075235] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215b2e3d-5668-4744-b47f-a7e6c9de7025 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.085054] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f4aa65-73bf-4571-a2b5-751f884f92f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.115656] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3e5da8-8a81-4d87-af1d-c3e9c7f88dd7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.123107] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80a1fa1-7904-4a7c-8360-c547e35291a5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.139400] env[61824]: DEBUG nova.compute.provider_tree [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.180640] env[61824]: DEBUG nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 981.397283] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275955, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.634256] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Successfully created port: 96cb8e2a-50c2-42c0-ac1a-7873710da8e2 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.643249] env[61824]: DEBUG nova.scheduler.client.report [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.710087] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.901511] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275955, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.011323} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.901867] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 8aa00861-bca6-4861-a5e5-0c538155c9ab/8aa00861-bca6-4861-a5e5-0c538155c9ab.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 981.902149] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.902458] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07556ae4-bc1e-4fe6-928e-c6e23b01cab8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.910713] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 981.910713] env[61824]: value = "task-1275956" [ 981.910713] env[61824]: _type = "Task" [ 981.910713] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.919183] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275956, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.955999] env[61824]: DEBUG nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 981.981845] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.982117] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.982284] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.982471] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.982622] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.982774] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.982997] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.983378] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.983568] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.983741] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.984046] env[61824]: DEBUG nova.virt.hardware [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.985739] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db581a75-c27f-4081-ab11-2dbafcb100bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.994128] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0252c79-18c0-4987-a698-e2d3f39000d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.150055] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.219s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.150055] env[61824]: DEBUG nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 982.152594] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.907s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.152814] env[61824]: DEBUG nova.objects.instance [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lazy-loading 'resources' on Instance uuid 82928174-1761-4296-bebe-7989c088d542 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.421985] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275956, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063872} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.423790] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.423790] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e511a73e-9a4f-48ad-8a02-96f1c2d742de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.446044] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 8aa00861-bca6-4861-a5e5-0c538155c9ab/8aa00861-bca6-4861-a5e5-0c538155c9ab.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.446241] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6839dbf-7c0b-4f39-9a3e-0d3cf1d6a9f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.466943] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 982.466943] env[61824]: value = "task-1275957" [ 982.466943] env[61824]: _type = "Task" [ 982.466943] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.474502] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275957, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.655396] env[61824]: DEBUG nova.compute.utils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.660089] env[61824]: DEBUG nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 982.660301] env[61824]: DEBUG nova.network.neutron [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 982.740973] env[61824]: DEBUG nova.policy [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8763f64818e24fad80ce232ac7d4463f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '629276280a234e7a8b810e6e74779fff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.796079] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3966e4-21f0-4d95-bf38-cdc373dd9ab6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.804269] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c097a797-630b-4f75-801f-e95ede17883b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.834271] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80230cde-4b01-4b3d-8cab-791ab5ef2e2e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.841328] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9515e225-307e-4463-a4d3-042c3ae26f8b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.855664] env[61824]: DEBUG nova.compute.provider_tree [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.977502] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275957, 'name': ReconfigVM_Task, 'duration_secs': 0.434783} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.977858] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 8aa00861-bca6-4861-a5e5-0c538155c9ab/8aa00861-bca6-4861-a5e5-0c538155c9ab.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.978527] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a352f77d-3389-46c7-ad37-64e4f330923c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.985915] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 982.985915] env[61824]: value = "task-1275958" [ 982.985915] env[61824]: _type = "Task" [ 982.985915] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.992361] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275958, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.020090] env[61824]: DEBUG nova.network.neutron [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Successfully created port: aa908ed5-c272-4dbe-8f6f-893a9044ab82 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.164368] env[61824]: DEBUG nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 983.267996] env[61824]: DEBUG nova.compute.manager [req-e43736d0-f267-495b-a291-eb448b076a37 req-6f914f5e-6cd0-489c-bfe7-9f20cb6bfe49 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received event network-vif-plugged-841a19df-b0cc-4ca3-a346-b9f56cbd1883 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.268316] env[61824]: DEBUG oslo_concurrency.lockutils [req-e43736d0-f267-495b-a291-eb448b076a37 req-6f914f5e-6cd0-489c-bfe7-9f20cb6bfe49 service nova] Acquiring lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.268525] env[61824]: DEBUG oslo_concurrency.lockutils [req-e43736d0-f267-495b-a291-eb448b076a37 req-6f914f5e-6cd0-489c-bfe7-9f20cb6bfe49 service nova] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.268696] env[61824]: DEBUG oslo_concurrency.lockutils [req-e43736d0-f267-495b-a291-eb448b076a37 req-6f914f5e-6cd0-489c-bfe7-9f20cb6bfe49 service nova] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.268866] env[61824]: DEBUG nova.compute.manager [req-e43736d0-f267-495b-a291-eb448b076a37 req-6f914f5e-6cd0-489c-bfe7-9f20cb6bfe49 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] No waiting events found dispatching network-vif-plugged-841a19df-b0cc-4ca3-a346-b9f56cbd1883 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.269072] env[61824]: WARNING nova.compute.manager [req-e43736d0-f267-495b-a291-eb448b076a37 req-6f914f5e-6cd0-489c-bfe7-9f20cb6bfe49 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received unexpected event network-vif-plugged-841a19df-b0cc-4ca3-a346-b9f56cbd1883 for instance with vm_state building and task_state spawning. [ 983.358535] env[61824]: DEBUG nova.scheduler.client.report [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.053862] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Successfully updated port: 841a19df-b0cc-4ca3-a346-b9f56cbd1883 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.058428] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.064167] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.782s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.064302] env[61824]: DEBUG nova.objects.instance [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'resources' on Instance uuid c5da84f7-4ee7-4537-a3f5-b95ee513c4dd {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.070970] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275958, 'name': Rename_Task, 'duration_secs': 0.132786} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.071242] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 984.071482] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d9a6a34-b33d-4e84-8432-ba49acbdd0aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.077537] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 984.077537] env[61824]: value = "task-1275959" [ 984.077537] env[61824]: _type = "Task" [ 984.077537] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.081205] env[61824]: INFO nova.scheduler.client.report [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Deleted allocations for instance 82928174-1761-4296-bebe-7989c088d542 [ 984.091830] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275959, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.564600] env[61824]: DEBUG nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 984.594813] env[61824]: DEBUG oslo_vmware.api [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1275959, 'name': PowerOnVM_Task, 'duration_secs': 0.427925} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.595275] env[61824]: DEBUG oslo_concurrency.lockutils [None req-dce56aaa-fd4e-429f-857c-b3b4fed94627 tempest-AttachInterfacesV270Test-1149035316 tempest-AttachInterfacesV270Test-1149035316-project-member] Lock "82928174-1761-4296-bebe-7989c088d542" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.366s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.602109] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.602386] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.602580] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.602738] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.602884] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.603053] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.603266] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.603424] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.603804] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.603804] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.603947] env[61824]: DEBUG nova.virt.hardware [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.604719] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.604719] env[61824]: INFO nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Took 8.31 seconds to spawn the instance on the hypervisor. [ 984.604719] env[61824]: DEBUG nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.610045] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8c68b0-bec9-4cbe-934c-205abf9ee8ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.620252] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88971986-b1c8-4a2d-890e-08d545af38f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.625670] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90eef1c0-65ae-413b-8093-5e05d3f464a8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.725879] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8d177e-229d-4717-8535-9fb497bf8e86 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.733719] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e898a1b-d22b-4c6a-a3e8-9e31a0763b31 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.765238] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb897ffc-dd57-4e2d-9b0d-9edad67fb9f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.773209] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0631f9-fb3e-416a-9730-5e33a95d6b13 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.790980] env[61824]: DEBUG nova.compute.provider_tree [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.832213] env[61824]: DEBUG nova.network.neutron [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Successfully updated port: aa908ed5-c272-4dbe-8f6f-893a9044ab82 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.149773] env[61824]: INFO nova.compute.manager [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Took 14.65 seconds to build instance. [ 985.293880] env[61824]: DEBUG nova.scheduler.client.report [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 985.302005] env[61824]: DEBUG nova.compute.manager [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received event network-changed-841a19df-b0cc-4ca3-a346-b9f56cbd1883 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.302509] env[61824]: DEBUG nova.compute.manager [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Refreshing instance network info cache due to event network-changed-841a19df-b0cc-4ca3-a346-b9f56cbd1883. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 985.302815] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Acquiring lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.303010] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Acquired lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.303232] env[61824]: DEBUG nova.network.neutron [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Refreshing network info cache for port 841a19df-b0cc-4ca3-a346-b9f56cbd1883 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 985.335206] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-ccb21ac1-6793-4cec-9c7e-824e5aecb488" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.335380] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-ccb21ac1-6793-4cec-9c7e-824e5aecb488" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.335603] env[61824]: DEBUG nova.network.neutron [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.481140] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.481559] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.481928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.482272] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.482609] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.486127] env[61824]: INFO nova.compute.manager [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Terminating instance [ 985.488824] env[61824]: DEBUG nova.compute.manager [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 985.489121] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 985.490916] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c7379b-8784-4877-87b0-f1b6923c7fdb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.501400] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 985.501639] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fff70c57-df28-45ca-8cc8-5e979a775409 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.509721] env[61824]: DEBUG oslo_vmware.api [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 985.509721] env[61824]: value = "task-1275960" [ 985.509721] env[61824]: _type = "Task" [ 985.509721] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.524799] env[61824]: DEBUG oslo_vmware.api [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.651635] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b5af805-cf70-49f2-a36a-d1e8cb2c6b0f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.166s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.800516] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.736s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.803720] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.093s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.805404] env[61824]: INFO nova.compute.claims [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 985.830849] env[61824]: INFO nova.scheduler.client.report [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted allocations for instance c5da84f7-4ee7-4537-a3f5-b95ee513c4dd [ 985.887872] env[61824]: DEBUG nova.network.neutron [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 985.923339] env[61824]: DEBUG nova.network.neutron [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 986.027151] env[61824]: DEBUG oslo_vmware.api [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275960, 'name': PowerOffVM_Task, 'duration_secs': 0.426452} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.028513] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.028513] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.028513] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3959fd13-dc9f-4537-a085-d3e01b2e189a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.090540] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.090747] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.091026] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleting the datastore file [datastore1] 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.091097] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a059670-14d2-4be0-964d-62f7984e5938 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.098419] env[61824]: DEBUG oslo_vmware.api [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for the task: (returnval){ [ 986.098419] env[61824]: value = "task-1275962" [ 986.098419] env[61824]: _type = "Task" [ 986.098419] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.108662] env[61824]: DEBUG oslo_vmware.api [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275962, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.122425] env[61824]: DEBUG nova.network.neutron [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.312617] env[61824]: DEBUG nova.network.neutron [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Updating instance_info_cache with network_info: [{"id": "aa908ed5-c272-4dbe-8f6f-893a9044ab82", "address": "fa:16:3e:8e:f5:43", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa908ed5-c2", "ovs_interfaceid": "aa908ed5-c272-4dbe-8f6f-893a9044ab82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.341396] env[61824]: DEBUG oslo_concurrency.lockutils [None req-573c1cfd-a0e0-4ca8-84ce-1c16e9a41d5c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "c5da84f7-4ee7-4537-a3f5-b95ee513c4dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.925s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.614506] env[61824]: DEBUG oslo_vmware.api [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Task: {'id': task-1275962, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131698} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.614616] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.614794] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 986.615192] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 986.615192] env[61824]: INFO nova.compute.manager [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 986.615510] env[61824]: DEBUG oslo.service.loopingcall [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.615641] env[61824]: DEBUG nova.compute.manager [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.615779] env[61824]: DEBUG nova.network.neutron [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 986.621218] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Successfully updated port: 96cb8e2a-50c2-42c0-ac1a-7873710da8e2 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.625019] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Releasing lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.625911] env[61824]: DEBUG nova.compute.manager [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Received event network-vif-plugged-aa908ed5-c272-4dbe-8f6f-893a9044ab82 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.625911] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Acquiring lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.625911] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.625911] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.625911] env[61824]: DEBUG nova.compute.manager [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] No waiting events found dispatching network-vif-plugged-aa908ed5-c272-4dbe-8f6f-893a9044ab82 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 986.626157] env[61824]: WARNING nova.compute.manager [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Received unexpected event network-vif-plugged-aa908ed5-c272-4dbe-8f6f-893a9044ab82 for instance with vm_state building and task_state spawning. [ 986.626202] env[61824]: DEBUG nova.compute.manager [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Received event network-changed-aa908ed5-c272-4dbe-8f6f-893a9044ab82 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 986.626359] env[61824]: DEBUG nova.compute.manager [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Refreshing instance network info cache due to event network-changed-aa908ed5-c272-4dbe-8f6f-893a9044ab82. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 986.626514] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Acquiring lock "refresh_cache-ccb21ac1-6793-4cec-9c7e-824e5aecb488" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.827825] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-ccb21ac1-6793-4cec-9c7e-824e5aecb488" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.828152] env[61824]: DEBUG nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Instance network_info: |[{"id": "aa908ed5-c272-4dbe-8f6f-893a9044ab82", "address": "fa:16:3e:8e:f5:43", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa908ed5-c2", "ovs_interfaceid": "aa908ed5-c272-4dbe-8f6f-893a9044ab82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 986.832138] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Acquired lock "refresh_cache-ccb21ac1-6793-4cec-9c7e-824e5aecb488" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.832403] env[61824]: DEBUG nova.network.neutron [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Refreshing network info cache for port aa908ed5-c272-4dbe-8f6f-893a9044ab82 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 986.834136] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:f5:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa908ed5-c272-4dbe-8f6f-893a9044ab82', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.843934] env[61824]: DEBUG oslo.service.loopingcall [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.850018] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 986.850270] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff53221d-a620-41be-96b9-dc0b6fc1ee47 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.875384] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.875384] env[61824]: value = "task-1275963" [ 986.875384] env[61824]: _type = "Task" [ 986.875384] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.888601] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275963, 'name': CreateVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.958823] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e83e820-0fdb-43b5-88ff-61150fc380e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.967204] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a1bb2a-3a5a-403a-b8a4-9233bece4f19 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.000251] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1a5c16-bafb-4a1d-87ee-c917aa7e4ef2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.008185] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d8b347-24c2-4410-9191-f4c7f35ceb17 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.022521] env[61824]: DEBUG nova.compute.provider_tree [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.124641] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.124641] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.124641] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.234668] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.234935] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.334456] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Received event network-changed-87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.334742] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Refreshing instance network info cache due to event network-changed-87ec4b1c-0d02-46ba-878d-b392ae74f472. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.334977] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Acquiring lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.335219] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Acquired lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.335491] env[61824]: DEBUG nova.network.neutron [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Refreshing network info cache for port 87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 987.351173] env[61824]: DEBUG nova.network.neutron [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.386053] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275963, 'name': CreateVM_Task, 'duration_secs': 0.306834} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.386181] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.386874] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.387080] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.387448] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.387648] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4f3be32-9361-4c55-b8bd-524ab5fa50cf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.392204] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 987.392204] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52457973-6dbc-194b-84c0-972cde55bc15" [ 987.392204] env[61824]: _type = "Task" [ 987.392204] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.401045] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52457973-6dbc-194b-84c0-972cde55bc15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.527873] env[61824]: DEBUG nova.scheduler.client.report [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.737625] env[61824]: DEBUG nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 987.811818] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 987.814562] env[61824]: DEBUG nova.network.neutron [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Updated VIF entry in instance network info cache for port aa908ed5-c272-4dbe-8f6f-893a9044ab82. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 987.814900] env[61824]: DEBUG nova.network.neutron [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Updating instance_info_cache with network_info: [{"id": "aa908ed5-c272-4dbe-8f6f-893a9044ab82", "address": "fa:16:3e:8e:f5:43", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa908ed5-c2", "ovs_interfaceid": "aa908ed5-c272-4dbe-8f6f-893a9044ab82", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.853749] env[61824]: INFO nova.compute.manager [-] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Took 1.24 seconds to deallocate network for instance. [ 987.903015] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52457973-6dbc-194b-84c0-972cde55bc15, 'name': SearchDatastore_Task, 'duration_secs': 0.008875} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.903317] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.903584] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.903789] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.904183] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.904183] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.904441] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bc3b7ea7-9549-4432-b33f-8f2a1b60fb49 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.914192] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.914384] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 987.915114] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11024d21-74ee-42f2-936c-54dc4066c7e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.923020] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 987.923020] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52fcafeb-71ad-e0a8-18ee-1cedcce3f405" [ 987.923020] env[61824]: _type = "Task" [ 987.923020] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.928250] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fcafeb-71ad-e0a8-18ee-1cedcce3f405, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.031083] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.031776] env[61824]: DEBUG nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.214574] env[61824]: DEBUG nova.network.neutron [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updated VIF entry in instance network info cache for port 87ec4b1c-0d02-46ba-878d-b392ae74f472. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 988.214950] env[61824]: DEBUG nova.network.neutron [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [{"id": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "address": "fa:16:3e:83:f8:c1", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ec4b1c-0d", "ovs_interfaceid": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.233709] env[61824]: DEBUG nova.network.neutron [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Updating instance_info_cache with network_info: [{"id": "841a19df-b0cc-4ca3-a346-b9f56cbd1883", "address": "fa:16:3e:11:c8:a4", "network": {"id": "2604cc15-2c0d-4433-a8f9-635f3a5fffee", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1941466340", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.173", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap841a19df-b0", "ovs_interfaceid": "841a19df-b0cc-4ca3-a346-b9f56cbd1883", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "96cb8e2a-50c2-42c0-ac1a-7873710da8e2", "address": "fa:16:3e:ce:70:62", "network": {"id": "0c04fe24-0160-4297-8354-1a70cf855c1b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1638025413", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.140", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52c1f5eb-3d4a-4faa-a30d-2b0a46430791", "external-id": "nsx-vlan-transportzone-775", "segmentation_id": 775, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96cb8e2a-50", "ovs_interfaceid": "96cb8e2a-50c2-42c0-ac1a-7873710da8e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.267031] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.267160] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.268661] env[61824]: INFO nova.compute.claims [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.317340] env[61824]: DEBUG oslo_concurrency.lockutils [req-3ede2920-943b-45f2-b1d3-361300556be7 req-8dd46b00-baf4-4db3-9be3-c2098500c7f4 service nova] Releasing lock "refresh_cache-ccb21ac1-6793-4cec-9c7e-824e5aecb488" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.362200] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.430949] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fcafeb-71ad-e0a8-18ee-1cedcce3f405, 'name': SearchDatastore_Task, 'duration_secs': 0.008682} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.431798] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ab4c850-eedc-447c-aa90-65737e1338d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.440447] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 988.440447] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52acb5dd-31d0-6741-5373-4d6976d430ea" [ 988.440447] env[61824]: _type = "Task" [ 988.440447] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.448139] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52acb5dd-31d0-6741-5373-4d6976d430ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.537163] env[61824]: DEBUG nova.compute.utils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.537770] env[61824]: DEBUG nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 988.537940] env[61824]: DEBUG nova.network.neutron [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 988.602081] env[61824]: DEBUG nova.policy [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2bcb8378feab42ae8666db25003481c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bd0e94b74d74f9898049c9cad364b5b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.718707] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Releasing lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.720691] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received event network-vif-plugged-96cb8e2a-50c2-42c0-ac1a-7873710da8e2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.720691] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Acquiring lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.720691] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.720691] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.720691] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] No waiting events found dispatching network-vif-plugged-96cb8e2a-50c2-42c0-ac1a-7873710da8e2 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 988.720691] env[61824]: WARNING nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received unexpected event network-vif-plugged-96cb8e2a-50c2-42c0-ac1a-7873710da8e2 for instance with vm_state building and task_state spawning. [ 988.720691] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received event network-changed-96cb8e2a-50c2-42c0-ac1a-7873710da8e2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.720691] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Refreshing instance network info cache due to event network-changed-96cb8e2a-50c2-42c0-ac1a-7873710da8e2. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.720691] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Acquiring lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.740271] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Releasing lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.740669] env[61824]: DEBUG nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Instance network_info: |[{"id": "841a19df-b0cc-4ca3-a346-b9f56cbd1883", "address": "fa:16:3e:11:c8:a4", "network": {"id": "2604cc15-2c0d-4433-a8f9-635f3a5fffee", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1941466340", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.173", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap841a19df-b0", "ovs_interfaceid": "841a19df-b0cc-4ca3-a346-b9f56cbd1883", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "96cb8e2a-50c2-42c0-ac1a-7873710da8e2", "address": "fa:16:3e:ce:70:62", "network": {"id": "0c04fe24-0160-4297-8354-1a70cf855c1b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1638025413", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.140", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52c1f5eb-3d4a-4faa-a30d-2b0a46430791", "external-id": "nsx-vlan-transportzone-775", "segmentation_id": 775, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96cb8e2a-50", "ovs_interfaceid": "96cb8e2a-50c2-42c0-ac1a-7873710da8e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 988.740993] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Acquired lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.741187] env[61824]: DEBUG nova.network.neutron [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Refreshing network info cache for port 96cb8e2a-50c2-42c0-ac1a-7873710da8e2 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.742516] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:c8:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1a9ee6f9-33be-4f58-8248-694024ec31d4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '841a19df-b0cc-4ca3-a346-b9f56cbd1883', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:70:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '52c1f5eb-3d4a-4faa-a30d-2b0a46430791', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96cb8e2a-50c2-42c0-ac1a-7873710da8e2', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.758288] env[61824]: DEBUG oslo.service.loopingcall [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.762217] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 988.762713] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6fb2e742-a14d-4058-8b12-c64df894b4a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.791291] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.791291] env[61824]: value = "task-1275964" [ 988.791291] env[61824]: _type = "Task" [ 988.791291] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.801880] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275964, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.956562] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52acb5dd-31d0-6741-5373-4d6976d430ea, 'name': SearchDatastore_Task, 'duration_secs': 0.00861} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.956858] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.957181] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] ccb21ac1-6793-4cec-9c7e-824e5aecb488/ccb21ac1-6793-4cec-9c7e-824e5aecb488.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 988.957301] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-41e02500-8ed3-48ff-ab9e-aef03c7863d2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.964883] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 988.964883] env[61824]: value = "task-1275965" [ 988.964883] env[61824]: _type = "Task" [ 988.964883] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.976163] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.041358] env[61824]: DEBUG nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.212689] env[61824]: DEBUG nova.network.neutron [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Successfully created port: 4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.313169] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275964, 'name': CreateVM_Task, 'duration_secs': 0.336315} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.313386] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.314302] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.314446] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.314847] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.320009] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b284b02d-7920-45e9-b88d-dfb47dc4a9f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.326600] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 989.326600] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525aa368-a6bf-aba7-1f16-abace891f473" [ 989.326600] env[61824]: _type = "Task" [ 989.326600] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.339658] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525aa368-a6bf-aba7-1f16-abace891f473, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.370023] env[61824]: DEBUG nova.network.neutron [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Updated VIF entry in instance network info cache for port 96cb8e2a-50c2-42c0-ac1a-7873710da8e2. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.370023] env[61824]: DEBUG nova.network.neutron [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Updating instance_info_cache with network_info: [{"id": "841a19df-b0cc-4ca3-a346-b9f56cbd1883", "address": "fa:16:3e:11:c8:a4", "network": {"id": "2604cc15-2c0d-4433-a8f9-635f3a5fffee", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1941466340", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.173", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1a9ee6f9-33be-4f58-8248-694024ec31d4", "external-id": "nsx-vlan-transportzone-581", "segmentation_id": 581, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap841a19df-b0", "ovs_interfaceid": "841a19df-b0cc-4ca3-a346-b9f56cbd1883", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "96cb8e2a-50c2-42c0-ac1a-7873710da8e2", "address": "fa:16:3e:ce:70:62", "network": {"id": "0c04fe24-0160-4297-8354-1a70cf855c1b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1638025413", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.140", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "773a259f132a42479464e120df314bb7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "52c1f5eb-3d4a-4faa-a30d-2b0a46430791", "external-id": "nsx-vlan-transportzone-775", "segmentation_id": 775, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96cb8e2a-50", "ovs_interfaceid": "96cb8e2a-50c2-42c0-ac1a-7873710da8e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.437219] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1085af67-f7b6-4540-9d68-f7cb80475bc0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.445187] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b5edab-c806-4db0-9841-96a1bcc23013 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.478370] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8927a41-a754-4132-bdb8-613f57ed9a65 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.487221] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466072} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.489530] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] ccb21ac1-6793-4cec-9c7e-824e5aecb488/ccb21ac1-6793-4cec-9c7e-824e5aecb488.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.490063] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.490372] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2033f530-c58c-4ea7-aac3-67f33ad269b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.493263] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6501971b-8cb2-4886-bd35-db0cc104f68e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.508309] env[61824]: DEBUG nova.compute.provider_tree [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.512015] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 989.512015] env[61824]: value = "task-1275966" [ 989.512015] env[61824]: _type = "Task" [ 989.512015] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.518924] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275966, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.847435] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525aa368-a6bf-aba7-1f16-abace891f473, 'name': SearchDatastore_Task, 'duration_secs': 0.049388} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.851517] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.851836] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 989.852095] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.852257] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.852451] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 989.852730] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66943cc0-cbb7-46c2-953e-228e409bfd20 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.862421] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 989.862641] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 989.863382] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc26a1d3-5109-48b4-9f63-9c366c036715 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.868881] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 989.868881] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c1a579-2369-7b06-5a8a-99a40571639d" [ 989.868881] env[61824]: _type = "Task" [ 989.868881] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.872178] env[61824]: DEBUG oslo_concurrency.lockutils [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] Releasing lock "refresh_cache-f29ed26b-94ed-43a5-baf0-84b6ba26fc22" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.872501] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Received event network-vif-deleted-45dc4201-9fb0-4060-ba24-f376e69a6e45 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.872611] env[61824]: INFO nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Neutron deleted interface 45dc4201-9fb0-4060-ba24-f376e69a6e45; detaching it from the instance and deleting it from the info cache [ 989.872776] env[61824]: DEBUG nova.network.neutron [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.878446] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c1a579-2369-7b06-5a8a-99a40571639d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.013038] env[61824]: DEBUG nova.scheduler.client.report [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.029628] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275966, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063174} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.029900] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.033787] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb69abd-3b60-4173-8690-ec8bbfe23290 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.057247] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] ccb21ac1-6793-4cec-9c7e-824e5aecb488/ccb21ac1-6793-4cec-9c7e-824e5aecb488.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.058373] env[61824]: DEBUG nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.060699] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f2dad36-d02e-401d-b3cc-70f6a360c503 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.080405] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 990.080405] env[61824]: value = "task-1275967" [ 990.080405] env[61824]: _type = "Task" [ 990.080405] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.088161] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275967, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.103696] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.103928] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.107050] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.107282] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.107441] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.107655] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.107814] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.107966] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.108189] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.108353] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.108519] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.108684] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.108858] env[61824]: DEBUG nova.virt.hardware [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.109903] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429da35f-f21e-49bd-b9a9-7d7f9d73d00d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.118058] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6314288-5301-4894-89af-f521da30937a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.375289] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06658154-b7ee-40a1-b9af-a1ee98ff63c2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.381306] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c1a579-2369-7b06-5a8a-99a40571639d, 'name': SearchDatastore_Task, 'duration_secs': 0.013702} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.382748] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15d87f9f-640a-4c64-b51d-42f1d577813b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.388729] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f516cf8f-344d-4b84-9fec-5db131c8e376 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.400678] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 990.400678] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]526a2069-38cc-cbfc-08b5-a058f9bb6054" [ 990.400678] env[61824]: _type = "Task" [ 990.400678] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.415108] env[61824]: DEBUG nova.compute.manager [req-3a4773d0-f8eb-4480-a008-9c75691c0184 req-936a266a-01f4-43b2-8be3-ee3df14ba815 service nova] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Detach interface failed, port_id=45dc4201-9fb0-4060-ba24-f376e69a6e45, reason: Instance 21759f11-80e9-419d-9f65-ca4767d1593b could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 990.418280] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526a2069-38cc-cbfc-08b5-a058f9bb6054, 'name': SearchDatastore_Task, 'duration_secs': 0.009675} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.418516] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.418763] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] f29ed26b-94ed-43a5-baf0-84b6ba26fc22/f29ed26b-94ed-43a5-baf0-84b6ba26fc22.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 990.419017] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12ffd4d1-b6e0-4398-a5ec-de7122a04153 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.425423] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 990.425423] env[61824]: value = "task-1275968" [ 990.425423] env[61824]: _type = "Task" [ 990.425423] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.432959] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275968, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.526027] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.526640] env[61824]: DEBUG nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.538058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.169s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.538058] env[61824]: DEBUG nova.objects.instance [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lazy-loading 'resources' on Instance uuid 21759f11-80e9-419d-9f65-ca4767d1593b {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.538058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.538058] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.590648] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275967, 'name': ReconfigVM_Task, 'duration_secs': 0.311986} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.590781] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Reconfigured VM instance instance-00000060 to attach disk [datastore2] ccb21ac1-6793-4cec-9c7e-824e5aecb488/ccb21ac1-6793-4cec-9c7e-824e5aecb488.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 990.591521] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f3b4fc0-bf7c-4a23-a0a3-a70af4c65edd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.598149] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 990.598149] env[61824]: value = "task-1275969" [ 990.598149] env[61824]: _type = "Task" [ 990.598149] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.613218] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275969, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.613218] env[61824]: DEBUG nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 990.860517] env[61824]: DEBUG nova.compute.manager [req-0b7ce570-653b-4a88-a650-cc9fe72f9f2a req-d6834c4c-f6a3-4d3f-9297-b8ded36107df service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-vif-plugged-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.860708] env[61824]: DEBUG oslo_concurrency.lockutils [req-0b7ce570-653b-4a88-a650-cc9fe72f9f2a req-d6834c4c-f6a3-4d3f-9297-b8ded36107df service nova] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.860937] env[61824]: DEBUG oslo_concurrency.lockutils [req-0b7ce570-653b-4a88-a650-cc9fe72f9f2a req-d6834c4c-f6a3-4d3f-9297-b8ded36107df service nova] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.861199] env[61824]: DEBUG oslo_concurrency.lockutils [req-0b7ce570-653b-4a88-a650-cc9fe72f9f2a req-d6834c4c-f6a3-4d3f-9297-b8ded36107df service nova] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.861388] env[61824]: DEBUG nova.compute.manager [req-0b7ce570-653b-4a88-a650-cc9fe72f9f2a req-d6834c4c-f6a3-4d3f-9297-b8ded36107df service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] No waiting events found dispatching network-vif-plugged-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 990.861563] env[61824]: WARNING nova.compute.manager [req-0b7ce570-653b-4a88-a650-cc9fe72f9f2a req-d6834c4c-f6a3-4d3f-9297-b8ded36107df service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received unexpected event network-vif-plugged-4c770083-4d53-46bb-b9a4-cdaa981b457c for instance with vm_state building and task_state spawning. [ 990.944917] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275968, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489113} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.944917] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] f29ed26b-94ed-43a5-baf0-84b6ba26fc22/f29ed26b-94ed-43a5-baf0-84b6ba26fc22.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 990.944917] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 990.944917] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d506c8cd-a3e5-4bd2-8094-4a2864da8e2c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.951019] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 990.951019] env[61824]: value = "task-1275970" [ 990.951019] env[61824]: _type = "Task" [ 990.951019] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.958469] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.974544] env[61824]: DEBUG nova.network.neutron [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Successfully updated port: 4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 991.032802] env[61824]: DEBUG nova.compute.utils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.034208] env[61824]: DEBUG nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.034383] env[61824]: DEBUG nova.network.neutron [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 991.040549] env[61824]: DEBUG nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 991.106354] env[61824]: DEBUG nova.policy [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1eb545ac2cf4f0d9ccddb040b17372b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e756c6817698464a9da391d4dd364946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.111545] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275969, 'name': Rename_Task, 'duration_secs': 0.178005} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.111813] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 991.112331] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f30361ee-b313-4324-a9a8-1e1b9662e501 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.123275] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 991.123275] env[61824]: value = "task-1275971" [ 991.123275] env[61824]: _type = "Task" [ 991.123275] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.138918] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275971, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.140512] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.184142] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716c1af3-d800-4ce9-be46-025618180121 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.191357] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862ba7ab-6702-4b2d-b322-b6708fc0352e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.221742] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4203200e-73d2-40cd-8f61-29e5b1e4a32c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.229090] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20878ce3-ae52-4c6c-ae0b-8901664b1153 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.243899] env[61824]: DEBUG nova.compute.provider_tree [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.347011] env[61824]: DEBUG nova.network.neutron [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Successfully created port: ba51fc90-3581-445e-b11f-a53f3c89d784 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.459524] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.476280] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.476430] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.476583] env[61824]: DEBUG nova.network.neutron [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.542390] env[61824]: DEBUG nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.565269] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.639789] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275971, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.747161] env[61824]: DEBUG nova.scheduler.client.report [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.960373] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.913215} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.960692] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 991.961513] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386d5621-7c75-4f4c-b81a-2f9142891cf6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.985258] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] f29ed26b-94ed-43a5-baf0-84b6ba26fc22/f29ed26b-94ed-43a5-baf0-84b6ba26fc22.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.987250] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e523c877-f079-4564-83ed-e5b3d9c9912a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.006897] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 992.006897] env[61824]: value = "task-1275972" [ 992.006897] env[61824]: _type = "Task" [ 992.006897] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.014854] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275972, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.028896] env[61824]: DEBUG nova.network.neutron [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 992.143029] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275971, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.253463] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.722s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.255910] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.115s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.257558] env[61824]: INFO nova.compute.claims [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.285262] env[61824]: INFO nova.scheduler.client.report [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Deleted allocations for instance 21759f11-80e9-419d-9f65-ca4767d1593b [ 992.322079] env[61824]: DEBUG nova.network.neutron [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c770083-4d", "ovs_interfaceid": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.516715] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275972, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.552177] env[61824]: DEBUG nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.576553] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.576810] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.576972] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.577179] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.577334] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.577483] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.577689] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.577850] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.578024] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.578190] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.578362] env[61824]: DEBUG nova.virt.hardware [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.579266] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413e83b1-efc2-4d79-864b-e1980a92bfac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.586829] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c6ddd7-f0ce-46b5-8c48-84eeeaac0ad1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.639308] env[61824]: DEBUG oslo_vmware.api [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275971, 'name': PowerOnVM_Task, 'duration_secs': 1.072665} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.639573] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.639780] env[61824]: INFO nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Took 8.07 seconds to spawn the instance on the hypervisor. [ 992.639962] env[61824]: DEBUG nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.640830] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbe1795-7a81-400e-b59e-28ac32c874a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.791656] env[61824]: DEBUG oslo_concurrency.lockutils [None req-370bf3ab-8116-434b-ba9a-666f097e5a1e tempest-ServersNegativeTestJSON-568153904 tempest-ServersNegativeTestJSON-568153904-project-member] Lock "21759f11-80e9-419d-9f65-ca4767d1593b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.310s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.826587] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.826587] env[61824]: DEBUG nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Instance network_info: |[{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c770083-4d", "ovs_interfaceid": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 992.826587] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:b8:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c770083-4d53-46bb-b9a4-cdaa981b457c', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.834277] env[61824]: DEBUG oslo.service.loopingcall [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.834277] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.834277] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5ab6fd9-5361-40c1-9a42-4b44abcbb8d2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.854245] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.854245] env[61824]: value = "task-1275973" [ 992.854245] env[61824]: _type = "Task" [ 992.854245] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.862268] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275973, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.893041] env[61824]: DEBUG nova.compute.manager [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.893260] env[61824]: DEBUG nova.compute.manager [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing instance network info cache due to event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.893480] env[61824]: DEBUG oslo_concurrency.lockutils [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] Acquiring lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.893628] env[61824]: DEBUG oslo_concurrency.lockutils [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] Acquired lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.893789] env[61824]: DEBUG nova.network.neutron [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 992.968213] env[61824]: DEBUG nova.network.neutron [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Successfully updated port: ba51fc90-3581-445e-b11f-a53f3c89d784 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.017733] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275972, 'name': ReconfigVM_Task, 'duration_secs': 0.659203} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.019091] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Reconfigured VM instance instance-0000005f to attach disk [datastore2] f29ed26b-94ed-43a5-baf0-84b6ba26fc22/f29ed26b-94ed-43a5-baf0-84b6ba26fc22.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.019091] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c7c8dd95-05d6-4c03-afe7-48859cb315b3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.024993] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 993.024993] env[61824]: value = "task-1275974" [ 993.024993] env[61824]: _type = "Task" [ 993.024993] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.032715] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275974, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.156694] env[61824]: INFO nova.compute.manager [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Took 17.62 seconds to build instance. [ 993.367641] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275973, 'name': CreateVM_Task, 'duration_secs': 0.500182} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.367941] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.368967] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.368967] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.368967] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 993.369986] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd7c6602-708c-4b2b-b3e7-1048a12e8e3a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.372168] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7943ac55-094e-4cac-9bfe-54d979f33583 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.377753] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 993.377753] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]527ed9f8-42e2-8798-b7e8-b582e1865311" [ 993.377753] env[61824]: _type = "Task" [ 993.377753] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.383179] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2077b7f2-04d5-4346-a9b8-8e8114827da7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.392169] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]527ed9f8-42e2-8798-b7e8-b582e1865311, 'name': SearchDatastore_Task, 'duration_secs': 0.010447} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.416153] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.416995] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.416995] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.416995] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.417179] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.419878] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c988bd36-f928-487f-b733-577c67992c37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.422248] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d15542-ed23-4c45-9370-24a2c3769b34 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.431339] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1275a54a-4698-427a-ab40-7dbbcca6d60d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.436280] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.436468] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.437552] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daca87b4-de51-4800-ac05-28d460e48919 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.448051] env[61824]: DEBUG nova.compute.provider_tree [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.452730] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 993.452730] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525cbef9-384b-283e-cfbe-b07c3927e6c2" [ 993.452730] env[61824]: _type = "Task" [ 993.452730] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.460858] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525cbef9-384b-283e-cfbe-b07c3927e6c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.469745] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "refresh_cache-1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.469889] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "refresh_cache-1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.470080] env[61824]: DEBUG nova.network.neutron [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 993.535088] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275974, 'name': Rename_Task, 'duration_secs': 0.238286} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.538127] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.538447] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-427cb60f-7571-4935-bc35-b43b3bb85604 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.545245] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 993.545245] env[61824]: value = "task-1275975" [ 993.545245] env[61824]: _type = "Task" [ 993.545245] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.555318] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275975, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.659431] env[61824]: DEBUG oslo_concurrency.lockutils [None req-95ef6014-46f0-406b-86d7-9e1799f5e4c6 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.130s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.660864] env[61824]: DEBUG nova.network.neutron [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updated VIF entry in instance network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 993.661423] env[61824]: DEBUG nova.network.neutron [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c770083-4d", "ovs_interfaceid": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.950983] env[61824]: DEBUG nova.scheduler.client.report [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.964600] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525cbef9-384b-283e-cfbe-b07c3927e6c2, 'name': SearchDatastore_Task, 'duration_secs': 0.010783} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.965435] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7451b05-4cb7-4744-bb05-825b21a7ab70 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.970978] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 993.970978] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52cbe225-8e72-3e12-7745-b17731f5e390" [ 993.970978] env[61824]: _type = "Task" [ 993.970978] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.981194] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52cbe225-8e72-3e12-7745-b17731f5e390, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.006735] env[61824]: DEBUG nova.network.neutron [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 994.057255] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275975, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.164789] env[61824]: DEBUG oslo_concurrency.lockutils [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] Releasing lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.165096] env[61824]: DEBUG nova.compute.manager [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Received event network-vif-plugged-ba51fc90-3581-445e-b11f-a53f3c89d784 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.165313] env[61824]: DEBUG oslo_concurrency.lockutils [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] Acquiring lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.165520] env[61824]: DEBUG oslo_concurrency.lockutils [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.165682] env[61824]: DEBUG oslo_concurrency.lockutils [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.165850] env[61824]: DEBUG nova.compute.manager [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] No waiting events found dispatching network-vif-plugged-ba51fc90-3581-445e-b11f-a53f3c89d784 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 994.166032] env[61824]: WARNING nova.compute.manager [req-8228bbd8-804a-4ed5-90d1-ca90b62bdd08 req-f517ed05-b8f4-44ad-aba0-e6182f2a0c2a service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Received unexpected event network-vif-plugged-ba51fc90-3581-445e-b11f-a53f3c89d784 for instance with vm_state building and task_state spawning. [ 994.240164] env[61824]: DEBUG nova.network.neutron [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Updating instance_info_cache with network_info: [{"id": "ba51fc90-3581-445e-b11f-a53f3c89d784", "address": "fa:16:3e:96:8f:28", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba51fc90-35", "ovs_interfaceid": "ba51fc90-3581-445e-b11f-a53f3c89d784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.463018] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.463018] env[61824]: DEBUG nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 994.463455] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.898s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.465414] env[61824]: INFO nova.compute.claims [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.483132] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52cbe225-8e72-3e12-7745-b17731f5e390, 'name': SearchDatastore_Task, 'duration_secs': 0.010476} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.483132] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.483462] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.483782] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e11e6456-b968-45a5-af9c-045ae056d9a7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.491210] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 994.491210] env[61824]: value = "task-1275976" [ 994.491210] env[61824]: _type = "Task" [ 994.491210] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.499899] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275976, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.556494] env[61824]: DEBUG oslo_vmware.api [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275975, 'name': PowerOnVM_Task, 'duration_secs': 0.539987} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.556764] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.556967] env[61824]: INFO nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Took 12.60 seconds to spawn the instance on the hypervisor. [ 994.557158] env[61824]: DEBUG nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 994.558143] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8bc9643-a3c2-41b6-b564-d322d5bb2237 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.677334] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.677732] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.747261] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "refresh_cache-1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.747261] env[61824]: DEBUG nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Instance network_info: |[{"id": "ba51fc90-3581-445e-b11f-a53f3c89d784", "address": "fa:16:3e:96:8f:28", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba51fc90-35", "ovs_interfaceid": "ba51fc90-3581-445e-b11f-a53f3c89d784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 994.747261] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:8f:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba51fc90-3581-445e-b11f-a53f3c89d784', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.758614] env[61824]: DEBUG oslo.service.loopingcall [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.759458] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 994.759740] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-849e7477-9fa1-44dd-a4ec-4a7a61b5c583 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.782351] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.782351] env[61824]: value = "task-1275977" [ 994.782351] env[61824]: _type = "Task" [ 994.782351] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.793782] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275977, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.971908] env[61824]: DEBUG nova.compute.utils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 994.973589] env[61824]: DEBUG nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 994.974095] env[61824]: DEBUG nova.network.neutron [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 995.001801] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275976, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448212} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.002105] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 995.002334] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.002703] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ea4b3d0-6389-4948-b050-69c11b6c0c22 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.009787] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 995.009787] env[61824]: value = "task-1275978" [ 995.009787] env[61824]: _type = "Task" [ 995.009787] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.018333] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275978, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.035448] env[61824]: DEBUG nova.policy [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd64f231608684672aed8882bf15faf7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9aff6d76e534482b3c4d98034a9591a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 995.078156] env[61824]: INFO nova.compute.manager [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Took 20.17 seconds to build instance. [ 995.145955] env[61824]: DEBUG nova.compute.manager [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Received event network-changed-ba51fc90-3581-445e-b11f-a53f3c89d784 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.146190] env[61824]: DEBUG nova.compute.manager [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Refreshing instance network info cache due to event network-changed-ba51fc90-3581-445e-b11f-a53f3c89d784. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 995.146423] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] Acquiring lock "refresh_cache-1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.146558] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] Acquired lock "refresh_cache-1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.146770] env[61824]: DEBUG nova.network.neutron [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Refreshing network info cache for port ba51fc90-3581-445e-b11f-a53f3c89d784 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 995.181477] env[61824]: DEBUG nova.compute.utils [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.292041] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275977, 'name': CreateVM_Task, 'duration_secs': 0.407545} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.292307] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 995.293018] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.293194] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.293521] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.293780] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e0f14d6-cba8-47dd-9b59-e8b1e8cca77a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.298392] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 995.298392] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5275adfe-0e9c-6f4f-3e73-05cf8f681bf3" [ 995.298392] env[61824]: _type = "Task" [ 995.298392] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.305965] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5275adfe-0e9c-6f4f-3e73-05cf8f681bf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.382186] env[61824]: DEBUG nova.network.neutron [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Successfully created port: 2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 995.477358] env[61824]: DEBUG nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 995.525356] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275978, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065651} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.525678] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.526464] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a93f18-fbd6-4632-84ee-ed3af7d8cc94 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.549160] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.549968] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecdcfb68-0751-45c7-ae12-1820e83d6833 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.571125] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 995.571125] env[61824]: value = "task-1275979" [ 995.571125] env[61824]: _type = "Task" [ 995.571125] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.578731] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275979, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.580252] env[61824]: DEBUG oslo_concurrency.lockutils [None req-863f9406-512c-4122-97e8-b799c3fbf88a tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.682s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.686377] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.690863] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f293fd-9c0d-4785-a92b-dee0dbf8d75c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.698467] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2579c192-0885-416e-b06e-9089282401d6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.730486] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0755a64-d061-483b-af9c-a68924a34e19 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.738407] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7e338a-8b53-4880-8c23-7f7647606caa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.752273] env[61824]: DEBUG nova.compute.provider_tree [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.810792] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5275adfe-0e9c-6f4f-3e73-05cf8f681bf3, 'name': SearchDatastore_Task, 'duration_secs': 0.009684} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.811722] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.811722] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.811722] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.811934] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.812187] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.812504] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a6e14e68-f728-438f-a4ee-f8bd70ce66dd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.820820] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.821078] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 995.821870] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4cc02fa-d845-4506-972a-ac310feee4f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.828715] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 995.828715] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525df5d0-7f00-464e-dfcb-fe980737aaf7" [ 995.828715] env[61824]: _type = "Task" [ 995.828715] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.836293] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525df5d0-7f00-464e-dfcb-fe980737aaf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.922539] env[61824]: DEBUG nova.network.neutron [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Updated VIF entry in instance network info cache for port ba51fc90-3581-445e-b11f-a53f3c89d784. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 995.923051] env[61824]: DEBUG nova.network.neutron [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Updating instance_info_cache with network_info: [{"id": "ba51fc90-3581-445e-b11f-a53f3c89d784", "address": "fa:16:3e:96:8f:28", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba51fc90-35", "ovs_interfaceid": "ba51fc90-3581-445e-b11f-a53f3c89d784", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.012561] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.012844] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.013076] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.013275] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.013453] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.015763] env[61824]: INFO nova.compute.manager [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Terminating instance [ 996.018434] env[61824]: DEBUG nova.compute.manager [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 996.018631] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.019664] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77fd85b-6844-4b63-a3ab-5fd3921d3841 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.029791] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.030052] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3e909f3-9bd6-4444-93dd-6a4954ec438c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.036013] env[61824]: DEBUG oslo_vmware.api [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 996.036013] env[61824]: value = "task-1275980" [ 996.036013] env[61824]: _type = "Task" [ 996.036013] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.045434] env[61824]: DEBUG oslo_vmware.api [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.080227] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275979, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.255742] env[61824]: DEBUG nova.scheduler.client.report [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.339405] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525df5d0-7f00-464e-dfcb-fe980737aaf7, 'name': SearchDatastore_Task, 'duration_secs': 0.00894} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.340251] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b57abd57-3cec-4ed6-aef8-eb03801c5b3d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.345969] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 996.345969] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5209c7df-bca9-d707-d594-f0fb581b2943" [ 996.345969] env[61824]: _type = "Task" [ 996.345969] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.353501] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5209c7df-bca9-d707-d594-f0fb581b2943, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.426425] env[61824]: DEBUG oslo_concurrency.lockutils [req-6ffcf69a-126c-4587-aae8-879c0e59bff4 req-899189dd-d390-492d-adf3-f2f517423882 service nova] Releasing lock "refresh_cache-1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.486558] env[61824]: DEBUG nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 996.510966] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 996.511266] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 996.511456] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 996.511661] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 996.511816] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 996.511968] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 996.512224] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 996.512401] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 996.512598] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 996.512775] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 996.512954] env[61824]: DEBUG nova.virt.hardware [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 996.513864] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d5d54b-65c3-4029-9b88-345496604b4d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.521604] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e398854-9c9a-446f-bf92-a3257eac9971 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.544314] env[61824]: DEBUG oslo_vmware.api [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275980, 'name': PowerOffVM_Task, 'duration_secs': 0.497202} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.544561] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 996.544730] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 996.544958] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-826a4115-e5f7-4df2-aaf4-2b50b04baff3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.579512] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275979, 'name': ReconfigVM_Task, 'duration_secs': 0.689717} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.579857] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.580398] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe0823bc-5120-460a-bead-0f610d62c7f6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.586970] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 996.586970] env[61824]: value = "task-1275982" [ 996.586970] env[61824]: _type = "Task" [ 996.586970] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.594896] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275982, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.659355] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 996.659608] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 996.659819] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Deleting the datastore file [datastore2] f29ed26b-94ed-43a5-baf0-84b6ba26fc22 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 996.660205] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1cf75d1-fc64-4613-adb6-27a846c79e8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.667116] env[61824]: DEBUG oslo_vmware.api [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for the task: (returnval){ [ 996.667116] env[61824]: value = "task-1275983" [ 996.667116] env[61824]: _type = "Task" [ 996.667116] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.674793] env[61824]: DEBUG oslo_vmware.api [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275983, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.764977] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.301s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.765666] env[61824]: DEBUG nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.768685] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.768965] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.769287] env[61824]: INFO nova.compute.manager [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Attaching volume f52be712-54ba-42fd-bfeb-86fa26dfb14f to /dev/sdb [ 996.803914] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb2ebdd-c189-49d0-b10e-1118279f9d87 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.816021] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc6d8ce-2dc7-4494-a8aa-03ab2733c138 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.827749] env[61824]: DEBUG nova.virt.block_device [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Updating existing volume attachment record: 25d91bc4-295a-4410-80e5-1953929d54ff {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 996.855476] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5209c7df-bca9-d707-d594-f0fb581b2943, 'name': SearchDatastore_Task, 'duration_secs': 0.008928} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.855816] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.856113] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d/1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 996.856398] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4eae307a-e8b0-46fd-9f70-d9352941c4f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.862151] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 996.862151] env[61824]: value = "task-1275984" [ 996.862151] env[61824]: _type = "Task" [ 996.862151] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.870016] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.099142] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275982, 'name': Rename_Task, 'duration_secs': 0.126571} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.099142] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.099142] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-22ed6485-d556-4001-bb5a-14ec81cacdb0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.104839] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 997.104839] env[61824]: value = "task-1275986" [ 997.104839] env[61824]: _type = "Task" [ 997.104839] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.112714] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275986, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.177813] env[61824]: DEBUG oslo_vmware.api [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Task: {'id': task-1275983, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165584} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.178149] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.178379] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.178588] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.178798] env[61824]: INFO nova.compute.manager [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Took 1.16 seconds to destroy the instance on the hypervisor. [ 997.179107] env[61824]: DEBUG oslo.service.loopingcall [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.179356] env[61824]: DEBUG nova.compute.manager [-] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 997.179515] env[61824]: DEBUG nova.network.neutron [-] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 997.190654] env[61824]: DEBUG nova.compute.manager [req-ae95ec2a-1a91-4308-9d51-e1a7816fdd8e req-e944b229-72b3-467b-bbdd-48a68ab8647c service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Received event network-vif-plugged-2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.190971] env[61824]: DEBUG oslo_concurrency.lockutils [req-ae95ec2a-1a91-4308-9d51-e1a7816fdd8e req-e944b229-72b3-467b-bbdd-48a68ab8647c service nova] Acquiring lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.191148] env[61824]: DEBUG oslo_concurrency.lockutils [req-ae95ec2a-1a91-4308-9d51-e1a7816fdd8e req-e944b229-72b3-467b-bbdd-48a68ab8647c service nova] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.191271] env[61824]: DEBUG oslo_concurrency.lockutils [req-ae95ec2a-1a91-4308-9d51-e1a7816fdd8e req-e944b229-72b3-467b-bbdd-48a68ab8647c service nova] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.191474] env[61824]: DEBUG nova.compute.manager [req-ae95ec2a-1a91-4308-9d51-e1a7816fdd8e req-e944b229-72b3-467b-bbdd-48a68ab8647c service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] No waiting events found dispatching network-vif-plugged-2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 997.191652] env[61824]: WARNING nova.compute.manager [req-ae95ec2a-1a91-4308-9d51-e1a7816fdd8e req-e944b229-72b3-467b-bbdd-48a68ab8647c service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Received unexpected event network-vif-plugged-2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 for instance with vm_state building and task_state spawning. [ 997.273269] env[61824]: DEBUG nova.compute.utils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.275059] env[61824]: DEBUG nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.275173] env[61824]: DEBUG nova.network.neutron [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 997.371994] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275984, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439826} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.372283] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d/1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 997.372504] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.372755] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a303d386-3da5-4b3a-ae36-35fc53713231 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.379189] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 997.379189] env[61824]: value = "task-1275989" [ 997.379189] env[61824]: _type = "Task" [ 997.379189] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.387081] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275989, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.409751] env[61824]: DEBUG nova.policy [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd64f231608684672aed8882bf15faf7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9aff6d76e534482b3c4d98034a9591a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 997.515986] env[61824]: DEBUG nova.network.neutron [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Successfully updated port: 2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.615850] env[61824]: DEBUG oslo_vmware.api [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1275986, 'name': PowerOnVM_Task, 'duration_secs': 0.505152} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.616317] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 997.616429] env[61824]: INFO nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Took 7.56 seconds to spawn the instance on the hypervisor. [ 997.616639] env[61824]: DEBUG nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.617611] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78cc9948-1354-4a9a-92ab-fa975d834b59 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.779237] env[61824]: DEBUG nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 997.889777] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275989, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061025} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.890082] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 997.890864] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9935f952-2fa8-494d-81a1-073e35d50757 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.913264] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d/1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.913589] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf4237de-57b1-4046-8d88-9db30bb8bee0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.929085] env[61824]: DEBUG nova.network.neutron [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Successfully created port: 3b69bd09-a533-499d-b63a-e125b6c295fd {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.936519] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 997.936519] env[61824]: value = "task-1275990" [ 997.936519] env[61824]: _type = "Task" [ 997.936519] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.945583] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275990, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.018480] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.018589] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.018736] env[61824]: DEBUG nova.network.neutron [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 998.135650] env[61824]: INFO nova.compute.manager [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Took 16.45 seconds to build instance. [ 998.438142] env[61824]: DEBUG nova.network.neutron [-] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.449295] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275990, 'name': ReconfigVM_Task, 'duration_secs': 0.268424} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.450252] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Reconfigured VM instance instance-00000062 to attach disk [datastore1] 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d/1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 998.450891] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5048bb62-3476-4d43-a7c2-bf8ef5ed3899 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.456945] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 998.456945] env[61824]: value = "task-1275991" [ 998.456945] env[61824]: _type = "Task" [ 998.456945] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.465136] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275991, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.552344] env[61824]: DEBUG nova.network.neutron [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 998.638834] env[61824]: DEBUG oslo_concurrency.lockutils [None req-ae5f5b62-f1b2-4ece-a64e-9519b7bc28ec tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.960s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.792143] env[61824]: DEBUG nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 998.817559] env[61824]: DEBUG nova.network.neutron [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Updating instance_info_cache with network_info: [{"id": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "address": "fa:16:3e:90:3d:72", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a20cdd6-6c", "ovs_interfaceid": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.822072] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.822328] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.822490] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.822673] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.822823] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.822976] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.823347] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.823527] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.823701] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.823896] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.824115] env[61824]: DEBUG nova.virt.hardware [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.824952] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd3a667-9deb-4b80-8373-881f774bb071 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.834111] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2de7d4-a358-4376-ac6a-a3fb59c635e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.944336] env[61824]: INFO nova.compute.manager [-] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Took 1.76 seconds to deallocate network for instance. [ 998.968156] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275991, 'name': Rename_Task, 'duration_secs': 0.143242} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.968611] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.968979] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-417995a5-8f52-416b-9622-e55e44f40ea8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.976097] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 998.976097] env[61824]: value = "task-1275992" [ 998.976097] env[61824]: _type = "Task" [ 998.976097] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.985973] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.222618] env[61824]: DEBUG nova.compute.manager [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Received event network-changed-2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.222949] env[61824]: DEBUG nova.compute.manager [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Refreshing instance network info cache due to event network-changed-2a20cdd6-6c0a-46d2-b5a7-1b78e2476561. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 999.223325] env[61824]: DEBUG oslo_concurrency.lockutils [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] Acquiring lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.236381] env[61824]: DEBUG nova.compute.manager [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.236728] env[61824]: DEBUG nova.compute.manager [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing instance network info cache due to event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 999.237097] env[61824]: DEBUG oslo_concurrency.lockutils [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] Acquiring lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.237097] env[61824]: DEBUG oslo_concurrency.lockutils [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] Acquired lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.237298] env[61824]: DEBUG nova.network.neutron [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 999.320644] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.321012] env[61824]: DEBUG nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Instance network_info: |[{"id": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "address": "fa:16:3e:90:3d:72", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a20cdd6-6c", "ovs_interfaceid": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 999.321372] env[61824]: DEBUG oslo_concurrency.lockutils [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] Acquired lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.321557] env[61824]: DEBUG nova.network.neutron [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Refreshing network info cache for port 2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 999.322817] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:3d:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a20cdd6-6c0a-46d2-b5a7-1b78e2476561', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.331193] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating folder: Project (f9aff6d76e534482b3c4d98034a9591a). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 999.335790] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-59f53707-db66-4e91-a24f-beca9d09b6a9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.346952] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created folder: Project (f9aff6d76e534482b3c4d98034a9591a) in parent group-v274074. [ 999.347343] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating folder: Instances. Parent ref: group-v274223. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 999.347630] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-757927ee-0cb5-42ed-8a4d-30ab02b356d9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.357463] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created folder: Instances in parent group-v274223. [ 999.357729] env[61824]: DEBUG oslo.service.loopingcall [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.357934] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 999.358162] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e093335a-252d-4104-8edf-d79197f2c854 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.380719] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.380719] env[61824]: value = "task-1275996" [ 999.380719] env[61824]: _type = "Task" [ 999.380719] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.388976] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275996, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.452014] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.452365] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.452727] env[61824]: DEBUG nova.objects.instance [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lazy-loading 'resources' on Instance uuid f29ed26b-94ed-43a5-baf0-84b6ba26fc22 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.485839] env[61824]: DEBUG oslo_vmware.api [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275992, 'name': PowerOnVM_Task, 'duration_secs': 0.461206} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.486157] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.486376] env[61824]: INFO nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Took 6.93 seconds to spawn the instance on the hypervisor. [ 999.486562] env[61824]: DEBUG nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 999.487410] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576ee9a2-2fd8-4c20-a183-562f7431c787 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.713386] env[61824]: DEBUG nova.network.neutron [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Updated VIF entry in instance network info cache for port 2a20cdd6-6c0a-46d2-b5a7-1b78e2476561. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 999.713764] env[61824]: DEBUG nova.network.neutron [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Updating instance_info_cache with network_info: [{"id": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "address": "fa:16:3e:90:3d:72", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a20cdd6-6c", "ovs_interfaceid": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.737597] env[61824]: DEBUG nova.network.neutron [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Successfully updated port: 3b69bd09-a533-499d-b63a-e125b6c295fd {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.894102] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275996, 'name': CreateVM_Task, 'duration_secs': 0.307381} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.894361] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 999.895300] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.895565] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.896319] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 999.896716] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0f24006-d746-407d-8858-07ad14bf7e8c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.903032] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 999.903032] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e88c84-10c5-44bc-0635-62e424bb36d4" [ 999.903032] env[61824]: _type = "Task" [ 999.903032] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.914629] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e88c84-10c5-44bc-0635-62e424bb36d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.003966] env[61824]: DEBUG nova.network.neutron [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updated VIF entry in instance network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1000.004345] env[61824]: DEBUG nova.network.neutron [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c770083-4d", "ovs_interfaceid": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.007502] env[61824]: INFO nova.compute.manager [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Took 11.77 seconds to build instance. [ 1000.071571] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6889df77-7c88-451d-99d5-90386c70048d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.078926] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d904f050-0eec-49f9-9c1f-5ef7cf2bc5b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.109594] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3248784c-5494-4c4e-ae01-0fa1a4dd1daa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.117351] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a04b1e6-534c-4c08-910f-2e8b8f9c2ea0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.131044] env[61824]: DEBUG nova.compute.provider_tree [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.217039] env[61824]: DEBUG oslo_concurrency.lockutils [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] Releasing lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.217441] env[61824]: DEBUG nova.compute.manager [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received event network-vif-deleted-96cb8e2a-50c2-42c0-ac1a-7873710da8e2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.217721] env[61824]: DEBUG nova.compute.manager [req-d6616411-9193-4c95-8b87-f88939f1c356 req-c9bf9a52-7135-4402-82b0-1fc579e5defb service nova] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Received event network-vif-deleted-841a19df-b0cc-4ca3-a346-b9f56cbd1883 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.239999] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.240401] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.240595] env[61824]: DEBUG nova.network.neutron [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.414544] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52e88c84-10c5-44bc-0635-62e424bb36d4, 'name': SearchDatastore_Task, 'duration_secs': 0.010527} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.415060] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.415169] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.415446] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.415535] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.415678] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.416310] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c470b6fa-415d-4c91-ad26-e7a0da9f42e0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.430278] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.430492] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.431260] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ab02aa9-54b9-4794-ab9c-606640cb2b77 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.436560] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1000.436560] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]526abda4-fa1b-934f-17f9-b9ffc1a86594" [ 1000.436560] env[61824]: _type = "Task" [ 1000.436560] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.444042] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526abda4-fa1b-934f-17f9-b9ffc1a86594, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.508899] env[61824]: DEBUG oslo_concurrency.lockutils [req-96ef2459-b7b0-4436-97d4-69760f401ec6 req-cb98a7aa-4036-4b0f-b759-b42ccebaf0aa service nova] Releasing lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.509841] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9ef0a642-c331-47f1-b501-617b9cd8513d tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.275s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.634166] env[61824]: DEBUG nova.scheduler.client.report [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.787082] env[61824]: DEBUG nova.network.neutron [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1000.932622] env[61824]: DEBUG nova.network.neutron [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Updating instance_info_cache with network_info: [{"id": "3b69bd09-a533-499d-b63a-e125b6c295fd", "address": "fa:16:3e:6a:c0:2f", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b69bd09-a5", "ovs_interfaceid": "3b69bd09-a533-499d-b63a-e125b6c295fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.950147] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526abda4-fa1b-934f-17f9-b9ffc1a86594, 'name': SearchDatastore_Task, 'duration_secs': 0.027132} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.950925] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1c207a1-f688-4e74-a0bd-6254b1f1c524 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.956626] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1000.956626] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52599f80-b72b-a189-0cc5-cc1525b9570a" [ 1000.956626] env[61824]: _type = "Task" [ 1000.956626] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.965587] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52599f80-b72b-a189-0cc5-cc1525b9570a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.139354] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.161414] env[61824]: INFO nova.scheduler.client.report [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Deleted allocations for instance f29ed26b-94ed-43a5-baf0-84b6ba26fc22 [ 1001.248561] env[61824]: DEBUG nova.compute.manager [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Received event network-vif-plugged-3b69bd09-a533-499d-b63a-e125b6c295fd {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.248561] env[61824]: DEBUG oslo_concurrency.lockutils [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] Acquiring lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.248703] env[61824]: DEBUG oslo_concurrency.lockutils [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.248886] env[61824]: DEBUG oslo_concurrency.lockutils [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.249179] env[61824]: DEBUG nova.compute.manager [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] No waiting events found dispatching network-vif-plugged-3b69bd09-a533-499d-b63a-e125b6c295fd {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.249415] env[61824]: WARNING nova.compute.manager [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Received unexpected event network-vif-plugged-3b69bd09-a533-499d-b63a-e125b6c295fd for instance with vm_state building and task_state spawning. [ 1001.249627] env[61824]: DEBUG nova.compute.manager [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Received event network-changed-3b69bd09-a533-499d-b63a-e125b6c295fd {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.249833] env[61824]: DEBUG nova.compute.manager [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Refreshing instance network info cache due to event network-changed-3b69bd09-a533-499d-b63a-e125b6c295fd. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1001.250028] env[61824]: DEBUG oslo_concurrency.lockutils [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] Acquiring lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.375480] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1001.375749] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274222', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'name': 'volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccb21ac1-6793-4cec-9c7e-824e5aecb488', 'attached_at': '', 'detached_at': '', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'serial': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1001.376647] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47ef74e-8678-40c0-b327-2244e27fddfc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.393368] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10295c11-05c4-4b1b-ac99-b88cc1ed4eca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.417251] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f/volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.417549] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af0d0648-eaaf-4c37-a682-513f9f953fd9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.430795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.431035] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.431239] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.431445] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.431634] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.434320] env[61824]: INFO nova.compute.manager [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Terminating instance [ 1001.435823] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.436102] env[61824]: DEBUG nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Instance network_info: |[{"id": "3b69bd09-a533-499d-b63a-e125b6c295fd", "address": "fa:16:3e:6a:c0:2f", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b69bd09-a5", "ovs_interfaceid": "3b69bd09-a533-499d-b63a-e125b6c295fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.436639] env[61824]: DEBUG nova.compute.manager [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1001.436830] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.437115] env[61824]: DEBUG oslo_concurrency.lockutils [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] Acquired lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.437288] env[61824]: DEBUG nova.network.neutron [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Refreshing network info cache for port 3b69bd09-a533-499d-b63a-e125b6c295fd {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1001.438351] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:c0:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b69bd09-a533-499d-b63a-e125b6c295fd', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.445878] env[61824]: DEBUG oslo.service.loopingcall [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.446522] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b75fd6-4c7f-407d-bf9d-4b6018117270 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.452587] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1001.453152] env[61824]: DEBUG oslo_vmware.api [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1001.453152] env[61824]: value = "task-1275997" [ 1001.453152] env[61824]: _type = "Task" [ 1001.453152] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.453348] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4ff3b82b-c9c2-440c-8e51-724ef7144872 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.477866] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.478182] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35071c5f-cf7a-4e9f-adb4-78b471b94ac0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.485937] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.485937] env[61824]: value = "task-1275998" [ 1001.485937] env[61824]: _type = "Task" [ 1001.485937] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.486518] env[61824]: DEBUG oslo_vmware.api [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275997, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.492792] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52599f80-b72b-a189-0cc5-cc1525b9570a, 'name': SearchDatastore_Task, 'duration_secs': 0.046318} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.493933] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.494225] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/4700de90-7230-43a5-87ae-f2c7ed9d4ea6.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1001.494531] env[61824]: DEBUG oslo_vmware.api [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1001.494531] env[61824]: value = "task-1275999" [ 1001.494531] env[61824]: _type = "Task" [ 1001.494531] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.495015] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-251839d5-d516-49e2-a8a4-9d7a72f4e26c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.502568] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275998, 'name': CreateVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.510044] env[61824]: DEBUG oslo_vmware.api [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275999, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.511302] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1001.511302] env[61824]: value = "task-1276000" [ 1001.511302] env[61824]: _type = "Task" [ 1001.511302] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.518821] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276000, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.669537] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6bf7e72d-6288-4754-9843-abefccef09c0 tempest-ServersTestMultiNic-2127385422 tempest-ServersTestMultiNic-2127385422-project-member] Lock "f29ed26b-94ed-43a5-baf0-84b6ba26fc22" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.657s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.691705] env[61824]: DEBUG nova.network.neutron [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Updated VIF entry in instance network info cache for port 3b69bd09-a533-499d-b63a-e125b6c295fd. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1001.692119] env[61824]: DEBUG nova.network.neutron [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Updating instance_info_cache with network_info: [{"id": "3b69bd09-a533-499d-b63a-e125b6c295fd", "address": "fa:16:3e:6a:c0:2f", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b69bd09-a5", "ovs_interfaceid": "3b69bd09-a533-499d-b63a-e125b6c295fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.979286] env[61824]: DEBUG oslo_vmware.api [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1275997, 'name': ReconfigVM_Task, 'duration_secs': 0.503591} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.979721] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Reconfigured VM instance instance-00000060 to attach disk [datastore1] volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f/volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.984434] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3d4f346-d8cc-4b9d-9130-f547168619e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.003848] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275998, 'name': CreateVM_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.005425] env[61824]: DEBUG oslo_vmware.api [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1002.005425] env[61824]: value = "task-1276001" [ 1002.005425] env[61824]: _type = "Task" [ 1002.005425] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.011074] env[61824]: DEBUG oslo_vmware.api [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1275999, 'name': PowerOffVM_Task, 'duration_secs': 0.222481} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.011819] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.011819] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.012107] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f4e4e16-39eb-4b7e-9e68-fdb1f3656dfd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.018525] env[61824]: DEBUG oslo_vmware.api [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276001, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.023928] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276000, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472619} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.024187] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/4700de90-7230-43a5-87ae-f2c7ed9d4ea6.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1002.024847] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1002.024847] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2816118-51a4-44f4-837a-50cd9247e282 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.029630] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1002.029630] env[61824]: value = "task-1276003" [ 1002.029630] env[61824]: _type = "Task" [ 1002.029630] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.037495] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276003, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.194901] env[61824]: DEBUG oslo_concurrency.lockutils [req-7c646b76-07e3-4e93-9f58-7eaee5e0cb7f req-f098e561-bf94-4dad-987b-79a9a4291371 service nova] Releasing lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.235171] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.235416] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.235603] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleting the datastore file [datastore1] 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.235872] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a04e7fe5-5607-4e21-9abd-27c3fb6a4894 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.243015] env[61824]: DEBUG oslo_vmware.api [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1002.243015] env[61824]: value = "task-1276004" [ 1002.243015] env[61824]: _type = "Task" [ 1002.243015] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.251874] env[61824]: DEBUG oslo_vmware.api [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276004, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.506915] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1275998, 'name': CreateVM_Task, 'duration_secs': 1.010697} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.509556] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1002.510245] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.510489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.511227] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1002.511596] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5aba6757-3865-4b73-b0cd-0cc4457ab1d7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.515868] env[61824]: DEBUG oslo_vmware.api [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276001, 'name': ReconfigVM_Task, 'duration_secs': 0.150951} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.516752] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274222', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'name': 'volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccb21ac1-6793-4cec-9c7e-824e5aecb488', 'attached_at': '', 'detached_at': '', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'serial': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1002.519163] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1002.519163] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52339ba2-41b3-291f-a662-6d848e909194" [ 1002.519163] env[61824]: _type = "Task" [ 1002.519163] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.527198] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52339ba2-41b3-291f-a662-6d848e909194, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.539963] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276003, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063615} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.540261] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1002.541024] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00098301-2eb8-4c49-9e49-a1dc3e116755 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.562399] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/4700de90-7230-43a5-87ae-f2c7ed9d4ea6.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.563121] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e3aa9bd-5aa1-4fbf-9813-bfce9e33a490 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.582798] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1002.582798] env[61824]: value = "task-1276005" [ 1002.582798] env[61824]: _type = "Task" [ 1002.582798] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.591523] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276005, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.752960] env[61824]: DEBUG oslo_vmware.api [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276004, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135514} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.753180] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.753394] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.753641] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.753758] env[61824]: INFO nova.compute.manager [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Took 1.32 seconds to destroy the instance on the hypervisor. [ 1002.753994] env[61824]: DEBUG oslo.service.loopingcall [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.754205] env[61824]: DEBUG nova.compute.manager [-] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1002.754339] env[61824]: DEBUG nova.network.neutron [-] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1003.033221] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52339ba2-41b3-291f-a662-6d848e909194, 'name': SearchDatastore_Task, 'duration_secs': 0.009093} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.033465] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.034266] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.034266] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.034266] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.034266] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.034419] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec2d80e3-ed40-4f87-a024-fd29c6c0d8a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.043373] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.043373] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.043373] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c434c4ec-2614-48b0-8048-48b14592bd8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.048076] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1003.048076] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5282501d-c0d1-9ade-8eca-eaf85b2a8f17" [ 1003.048076] env[61824]: _type = "Task" [ 1003.048076] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.055411] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5282501d-c0d1-9ade-8eca-eaf85b2a8f17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.093847] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276005, 'name': ReconfigVM_Task, 'duration_secs': 0.243009} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.094151] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/4700de90-7230-43a5-87ae-f2c7ed9d4ea6.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.094789] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eeb99c4c-e60f-4fc5-8849-e93c5bb9fa13 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.101646] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1003.101646] env[61824]: value = "task-1276006" [ 1003.101646] env[61824]: _type = "Task" [ 1003.101646] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.109564] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276006, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.285249] env[61824]: DEBUG nova.compute.manager [req-dd343258-ae96-4402-b12c-7856b9642ac2 req-b19217da-b709-42c8-a350-c8044e1c538f service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Received event network-vif-deleted-ba51fc90-3581-445e-b11f-a53f3c89d784 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.285459] env[61824]: INFO nova.compute.manager [req-dd343258-ae96-4402-b12c-7856b9642ac2 req-b19217da-b709-42c8-a350-c8044e1c538f service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Neutron deleted interface ba51fc90-3581-445e-b11f-a53f3c89d784; detaching it from the instance and deleting it from the info cache [ 1003.285637] env[61824]: DEBUG nova.network.neutron [req-dd343258-ae96-4402-b12c-7856b9642ac2 req-b19217da-b709-42c8-a350-c8044e1c538f service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.484471] env[61824]: DEBUG nova.network.neutron [-] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.559852] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5282501d-c0d1-9ade-8eca-eaf85b2a8f17, 'name': SearchDatastore_Task, 'duration_secs': 0.009039} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.560739] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb5f88eb-deac-496a-9d16-9a28f1712791 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.563489] env[61824]: DEBUG nova.objects.instance [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'flavor' on Instance uuid ccb21ac1-6793-4cec-9c7e-824e5aecb488 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.568200] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1003.568200] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]529b1d72-68cf-ab04-603c-8c0e97983ad4" [ 1003.568200] env[61824]: _type = "Task" [ 1003.568200] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.576951] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529b1d72-68cf-ab04-603c-8c0e97983ad4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.611627] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276006, 'name': Rename_Task, 'duration_secs': 0.122117} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.612063] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1003.612343] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b5fdb19-9a0a-48fb-aead-a2328dcecf28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.619084] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1003.619084] env[61824]: value = "task-1276007" [ 1003.619084] env[61824]: _type = "Task" [ 1003.619084] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.627205] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276007, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.792511] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dca2fca-0eb8-4d53-9ffd-90563117bb00 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.802204] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7d04e3-f858-4941-840e-0d2032468bba {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.829534] env[61824]: DEBUG nova.compute.manager [req-dd343258-ae96-4402-b12c-7856b9642ac2 req-b19217da-b709-42c8-a350-c8044e1c538f service nova] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Detach interface failed, port_id=ba51fc90-3581-445e-b11f-a53f3c89d784, reason: Instance 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1003.986746] env[61824]: INFO nova.compute.manager [-] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Took 1.23 seconds to deallocate network for instance. [ 1004.068043] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a3902ca2-76d1-4d22-b447-fcada5cd2519 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.299s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.078809] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529b1d72-68cf-ab04-603c-8c0e97983ad4, 'name': SearchDatastore_Task, 'duration_secs': 0.009081} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.078809] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.078988] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/1ddb6d7a-d8e1-492d-84c9-d61e27faa105.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.079201] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-04eae63d-b15f-44d5-824c-80f041918d6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.085471] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1004.085471] env[61824]: value = "task-1276008" [ 1004.085471] env[61824]: _type = "Task" [ 1004.085471] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.093041] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.127950] env[61824]: DEBUG oslo_vmware.api [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276007, 'name': PowerOnVM_Task, 'duration_secs': 0.419826} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.128232] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1004.128466] env[61824]: INFO nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Took 7.64 seconds to spawn the instance on the hypervisor. [ 1004.128694] env[61824]: DEBUG nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1004.129515] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb883080-5ae7-4758-8f70-22533594388e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.152635] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.153158] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.153406] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.153640] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.153868] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.155985] env[61824]: INFO nova.compute.manager [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Terminating instance [ 1004.158227] env[61824]: DEBUG nova.compute.manager [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.158342] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.158596] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3754914-2837-4ed8-be2a-dc28d033180a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.165594] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1004.165594] env[61824]: value = "task-1276009" [ 1004.165594] env[61824]: _type = "Task" [ 1004.165594] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.174320] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276009, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.494672] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.495091] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.495347] env[61824]: DEBUG nova.objects.instance [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'resources' on Instance uuid 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1004.595910] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463664} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.596212] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/1ddb6d7a-d8e1-492d-84c9-d61e27faa105.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.596432] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.596681] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65868d71-0eea-49ec-9148-fabcf1893ecb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.602808] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1004.602808] env[61824]: value = "task-1276010" [ 1004.602808] env[61824]: _type = "Task" [ 1004.602808] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.609929] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276010, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.644466] env[61824]: INFO nova.compute.manager [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Took 13.53 seconds to build instance. [ 1004.675299] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276009, 'name': PowerOffVM_Task, 'duration_secs': 0.16757} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.675550] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.675748] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1004.675938] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274222', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'name': 'volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccb21ac1-6793-4cec-9c7e-824e5aecb488', 'attached_at': '', 'detached_at': '', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'serial': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1004.676728] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b450d12e-d689-43a9-8471-b6b100dc2215 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.697706] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49535df9-04eb-4f3d-a6e3-b5ef26c13785 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.704011] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0306531-19c9-4045-8cd4-a1901ace3728 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.723202] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a967ff0-708c-4361-9531-aca1974a68cb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.737081] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] The volume has not been displaced from its original location: [datastore1] volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f/volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1004.742267] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1004.742519] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f94ab8af-e213-4f26-955e-3dea995385bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.759455] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1004.759455] env[61824]: value = "task-1276011" [ 1004.759455] env[61824]: _type = "Task" [ 1004.759455] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.766512] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276011, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.103450] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065e860c-a05f-4bd2-96df-879077977f56 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.116649] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276010, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.119020] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723200e0-6610-4d48-96d1-c4f2d5c5d171 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.148535] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cf1c9bac-65cd-48f2-beda-f18c92775f86 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.044s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.149662] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab802a5e-dd1c-4e3a-b692-61bbffba5d35 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.157198] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f051a9c-1831-43ce-92c3-7cfe60fbdb8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.170752] env[61824]: DEBUG nova.compute.provider_tree [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.269869] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276011, 'name': ReconfigVM_Task, 'duration_secs': 0.240631} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.270209] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1005.274941] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04385dee-6c6d-40ad-9925-81f0c63abc09 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.290380] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1005.290380] env[61824]: value = "task-1276012" [ 1005.290380] env[61824]: _type = "Task" [ 1005.290380] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.298343] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276012, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.616695] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276010, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.674277] env[61824]: DEBUG nova.scheduler.client.report [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1005.800679] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276012, 'name': ReconfigVM_Task, 'duration_secs': 0.466498} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.801010] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274222', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'name': 'volume-f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ccb21ac1-6793-4cec-9c7e-824e5aecb488', 'attached_at': '', 'detached_at': '', 'volume_id': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f', 'serial': 'f52be712-54ba-42fd-bfeb-86fa26dfb14f'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1005.801320] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1005.802094] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeea7e97-9d43-44eb-b041-05e7f8eb7e32 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.808482] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1005.808715] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e9e119e-c1e4-411f-850b-3a4b8905b2ea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.871237] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1005.871502] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1005.871703] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleting the datastore file [datastore2] ccb21ac1-6793-4cec-9c7e-824e5aecb488 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.871979] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fe7e97a-9ee9-4a3a-9238-a234b4f952f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.880343] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1005.880343] env[61824]: value = "task-1276014" [ 1005.880343] env[61824]: _type = "Task" [ 1005.880343] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.887970] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.116587] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276010, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.290906} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.116924] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1006.117659] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef90514-66c9-4638-b04d-aaad50e8d1a0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.139146] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/1ddb6d7a-d8e1-492d-84c9-d61e27faa105.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.139485] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33a50d2b-f39c-43e8-9345-217a04056f6f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.158700] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1006.158700] env[61824]: value = "task-1276015" [ 1006.158700] env[61824]: _type = "Task" [ 1006.158700] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.166543] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276015, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.179638] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.205854] env[61824]: INFO nova.scheduler.client.report [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted allocations for instance 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d [ 1006.391274] env[61824]: DEBUG oslo_vmware.api [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.421504} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.391574] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.391770] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1006.391952] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1006.392146] env[61824]: INFO nova.compute.manager [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1006.392391] env[61824]: DEBUG oslo.service.loopingcall [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.392585] env[61824]: DEBUG nova.compute.manager [-] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.392679] env[61824]: DEBUG nova.network.neutron [-] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1006.665872] env[61824]: DEBUG nova.compute.manager [req-da3f44cd-49b0-4f86-8eb9-af3408601216 req-2f6923b0-b997-4d8b-bac1-6f160f01df7c service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Received event network-vif-deleted-aa908ed5-c272-4dbe-8f6f-893a9044ab82 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.666172] env[61824]: INFO nova.compute.manager [req-da3f44cd-49b0-4f86-8eb9-af3408601216 req-2f6923b0-b997-4d8b-bac1-6f160f01df7c service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Neutron deleted interface aa908ed5-c272-4dbe-8f6f-893a9044ab82; detaching it from the instance and deleting it from the info cache [ 1006.666411] env[61824]: DEBUG nova.network.neutron [req-da3f44cd-49b0-4f86-8eb9-af3408601216 req-2f6923b0-b997-4d8b-bac1-6f160f01df7c service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.674230] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276015, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.715397] env[61824]: DEBUG oslo_concurrency.lockutils [None req-efa7a5a2-a843-4c4d-afe2-270a6fe3f448 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.284s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.138154] env[61824]: DEBUG nova.network.neutron [-] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.169257] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276015, 'name': ReconfigVM_Task, 'duration_secs': 0.532367} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.169503] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5e67007-7a03-4edc-a86e-8f22064964cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.171304] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/1ddb6d7a-d8e1-492d-84c9-d61e27faa105.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.171967] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-887d375a-27fc-4c31-8bb8-b1ff782c0639 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.182518] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f137a23-131b-4e3a-8fcb-3f0ff20b28c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.192916] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1007.192916] env[61824]: value = "task-1276016" [ 1007.192916] env[61824]: _type = "Task" [ 1007.192916] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.204054] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276016, 'name': Rename_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.211228] env[61824]: DEBUG nova.compute.manager [req-da3f44cd-49b0-4f86-8eb9-af3408601216 req-2f6923b0-b997-4d8b-bac1-6f160f01df7c service nova] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Detach interface failed, port_id=aa908ed5-c272-4dbe-8f6f-893a9044ab82, reason: Instance ccb21ac1-6793-4cec-9c7e-824e5aecb488 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1007.641453] env[61824]: INFO nova.compute.manager [-] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Took 1.25 seconds to deallocate network for instance. [ 1007.704061] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276016, 'name': Rename_Task, 'duration_secs': 0.410926} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.704365] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.704610] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42f0ba7a-2d9b-47ca-a86f-8adc515d1e40 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.710742] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1007.710742] env[61824]: value = "task-1276017" [ 1007.710742] env[61824]: _type = "Task" [ 1007.710742] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.717849] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276017, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.787746] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.787991] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.183356] env[61824]: INFO nova.compute.manager [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Took 0.54 seconds to detach 1 volumes for instance. [ 1008.221320] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276017, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.290380] env[61824]: DEBUG nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.690031] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.690310] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.690508] env[61824]: DEBUG nova.objects.instance [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'resources' on Instance uuid ccb21ac1-6793-4cec-9c7e-824e5aecb488 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.721129] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276017, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.812442] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.222367] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276017, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.279548] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd42c9f-b2f6-4928-8acd-1766b40b7069 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.286683] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58114f8c-5a3b-476a-a4e1-9d7f35e8a9d0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.317293] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a062d5b3-b371-4eec-a0a8-fc9152090e1b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.324240] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8093f56b-605b-42c1-aee9-80fef315488d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.336991] env[61824]: DEBUG nova.compute.provider_tree [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.723066] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276017, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.840484] env[61824]: DEBUG nova.scheduler.client.report [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.223264] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276017, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.345500] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.655s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.347952] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.535s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.349829] env[61824]: INFO nova.compute.claims [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.363903] env[61824]: INFO nova.scheduler.client.report [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted allocations for instance ccb21ac1-6793-4cec-9c7e-824e5aecb488 [ 1010.723670] env[61824]: DEBUG oslo_vmware.api [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276017, 'name': PowerOnVM_Task, 'duration_secs': 2.727503} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.723940] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1010.724176] env[61824]: INFO nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Took 11.93 seconds to spawn the instance on the hypervisor. [ 1010.724368] env[61824]: DEBUG nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1010.725223] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9e56a6-e6cc-4f6c-8cb8-d63f6b085f55 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.871399] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a28acda-0ba1-4526-9608-5389cd6d7931 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "ccb21ac1-6793-4cec-9c7e-824e5aecb488" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.718s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.240873] env[61824]: INFO nova.compute.manager [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Took 19.69 seconds to build instance. [ 1011.435352] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158acaa0-8fb3-43e7-87bd-fbd7289b2116 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.443287] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df70944c-3dc8-4599-8636-ccbea9c2ce62 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.474295] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3e7987-078a-4d2e-a3a8-5775c3d8da73 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.481965] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa0f38a-276f-4ad6-8e20-50b41a6b82c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.495408] env[61824]: DEBUG nova.compute.provider_tree [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.742113] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d0f1aef2-6340-4baa-9b86-c44737cc6cf7 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.209s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.998065] env[61824]: DEBUG nova.scheduler.client.report [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.109670] env[61824]: INFO nova.compute.manager [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Rescuing [ 1012.109987] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.110169] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.110370] env[61824]: DEBUG nova.network.neutron [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1012.505777] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.158s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.506525] env[61824]: DEBUG nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.785820] env[61824]: DEBUG nova.network.neutron [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Updating instance_info_cache with network_info: [{"id": "3b69bd09-a533-499d-b63a-e125b6c295fd", "address": "fa:16:3e:6a:c0:2f", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b69bd09-a5", "ovs_interfaceid": "3b69bd09-a533-499d-b63a-e125b6c295fd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.919795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "76303768-1bb1-4a82-8ca8-df301c19d02d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.920108] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.012140] env[61824]: DEBUG nova.compute.utils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.013692] env[61824]: DEBUG nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1013.013918] env[61824]: DEBUG nova.network.neutron [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1013.060491] env[61824]: DEBUG nova.policy [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1eb545ac2cf4f0d9ccddb040b17372b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e756c6817698464a9da391d4dd364946', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1013.290642] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-1ddb6d7a-d8e1-492d-84c9-d61e27faa105" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.305922] env[61824]: DEBUG nova.network.neutron [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Successfully created port: d998ac38-5777-435b-96bc-d96f4bbcb6b4 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1013.424836] env[61824]: DEBUG nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1013.516818] env[61824]: DEBUG nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.828984] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.829312] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2b3bb78-2513-42b4-9c42-98fd46036433 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.836692] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1013.836692] env[61824]: value = "task-1276018" [ 1013.836692] env[61824]: _type = "Task" [ 1013.836692] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.844686] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.945077] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.945989] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.946942] env[61824]: INFO nova.compute.claims [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.347018] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276018, 'name': PowerOffVM_Task, 'duration_secs': 0.1789} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.347315] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.348097] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a69af10-7a17-4486-9bce-fcaf3946b21b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.366841] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d83443-ec20-4bf1-ae17-e5488784e0be {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.395696] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.395974] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8188558-0680-4d05-bcfc-52247af7720c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.401965] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1014.401965] env[61824]: value = "task-1276019" [ 1014.401965] env[61824]: _type = "Task" [ 1014.401965] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.409439] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276019, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.526388] env[61824]: DEBUG nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.547367] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.547632] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.547797] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.547986] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.548157] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.548311] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.548521] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.548685] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.548855] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.549033] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.549229] env[61824]: DEBUG nova.virt.hardware [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.550110] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a963e2-881d-40db-a0f4-662d6bc59895 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.557706] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ee29ee-d2fd-4cc4-a54d-1f2ecf0e3ea5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.739719] env[61824]: DEBUG nova.compute.manager [req-21deb44b-7839-461a-a94c-302edc4cb945 req-d4220ea3-1b5f-45f5-ad3f-8b10eca8c5c7 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Received event network-vif-plugged-d998ac38-5777-435b-96bc-d96f4bbcb6b4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.739929] env[61824]: DEBUG oslo_concurrency.lockutils [req-21deb44b-7839-461a-a94c-302edc4cb945 req-d4220ea3-1b5f-45f5-ad3f-8b10eca8c5c7 service nova] Acquiring lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.740164] env[61824]: DEBUG oslo_concurrency.lockutils [req-21deb44b-7839-461a-a94c-302edc4cb945 req-d4220ea3-1b5f-45f5-ad3f-8b10eca8c5c7 service nova] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.740342] env[61824]: DEBUG oslo_concurrency.lockutils [req-21deb44b-7839-461a-a94c-302edc4cb945 req-d4220ea3-1b5f-45f5-ad3f-8b10eca8c5c7 service nova] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.740515] env[61824]: DEBUG nova.compute.manager [req-21deb44b-7839-461a-a94c-302edc4cb945 req-d4220ea3-1b5f-45f5-ad3f-8b10eca8c5c7 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] No waiting events found dispatching network-vif-plugged-d998ac38-5777-435b-96bc-d96f4bbcb6b4 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.740684] env[61824]: WARNING nova.compute.manager [req-21deb44b-7839-461a-a94c-302edc4cb945 req-d4220ea3-1b5f-45f5-ad3f-8b10eca8c5c7 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Received unexpected event network-vif-plugged-d998ac38-5777-435b-96bc-d96f4bbcb6b4 for instance with vm_state building and task_state spawning. [ 1014.912318] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1014.912578] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.912823] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.912978] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.913177] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.913435] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2410f67-8fe7-4a3e-93bf-a28480e2fc70 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.922198] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.922376] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.923067] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bddaddd-3113-41f1-b759-d382fb9d142c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.927632] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1014.927632] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525e99c0-38cd-cb6e-b0c2-241f0fe6742f" [ 1014.927632] env[61824]: _type = "Task" [ 1014.927632] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.935103] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525e99c0-38cd-cb6e-b0c2-241f0fe6742f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.044255] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-601599d7-8c3a-4b5b-a75a-0a85ffc4d1ff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.051732] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77b2569-cda4-46d9-b0ed-a47495f71e6a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.082369] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001a07fd-b195-4712-b62a-897464bdaa84 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.089505] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55a0e3a-f415-41e8-b383-1ac6011c928d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.102444] env[61824]: DEBUG nova.compute.provider_tree [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.303365] env[61824]: DEBUG nova.network.neutron [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Successfully updated port: d998ac38-5777-435b-96bc-d96f4bbcb6b4 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1015.326459] env[61824]: DEBUG nova.compute.manager [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Received event network-changed-d998ac38-5777-435b-96bc-d96f4bbcb6b4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.326702] env[61824]: DEBUG nova.compute.manager [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Refreshing instance network info cache due to event network-changed-d998ac38-5777-435b-96bc-d96f4bbcb6b4. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.326877] env[61824]: DEBUG oslo_concurrency.lockutils [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] Acquiring lock "refresh_cache-d97da7c1-d20d-45db-a03d-c9d4ceb8c235" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.327027] env[61824]: DEBUG oslo_concurrency.lockutils [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] Acquired lock "refresh_cache-d97da7c1-d20d-45db-a03d-c9d4ceb8c235" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.327201] env[61824]: DEBUG nova.network.neutron [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Refreshing network info cache for port d998ac38-5777-435b-96bc-d96f4bbcb6b4 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1015.438370] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525e99c0-38cd-cb6e-b0c2-241f0fe6742f, 'name': SearchDatastore_Task, 'duration_secs': 0.013895} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.439146] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ce39ea2-2fc1-42d6-b7f4-06e7b93b6308 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.444387] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1015.444387] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52695768-dea4-fa80-7855-9c93aaa76396" [ 1015.444387] env[61824]: _type = "Task" [ 1015.444387] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.452168] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52695768-dea4-fa80-7855-9c93aaa76396, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.605254] env[61824]: DEBUG nova.scheduler.client.report [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.805693] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "refresh_cache-d97da7c1-d20d-45db-a03d-c9d4ceb8c235" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.865171] env[61824]: DEBUG nova.network.neutron [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1015.956347] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52695768-dea4-fa80-7855-9c93aaa76396, 'name': SearchDatastore_Task, 'duration_secs': 0.009578} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.956637] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.956919] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk. {{(pid=61824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1015.957209] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69d852db-063f-4bf0-81fc-88fc707099fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.964212] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1015.964212] env[61824]: value = "task-1276020" [ 1015.964212] env[61824]: _type = "Task" [ 1015.964212] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.964991] env[61824]: DEBUG nova.network.neutron [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.973693] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276020, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.110918] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.165s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.111600] env[61824]: DEBUG nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.472354] env[61824]: DEBUG oslo_concurrency.lockutils [req-20d466a9-d829-4720-8e89-4ce885291169 req-32f10122-7d34-4d1c-93e5-bfdb2b0bd758 service nova] Releasing lock "refresh_cache-d97da7c1-d20d-45db-a03d-c9d4ceb8c235" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.472653] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "refresh_cache-d97da7c1-d20d-45db-a03d-c9d4ceb8c235" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.472818] env[61824]: DEBUG nova.network.neutron [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.479181] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276020, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460242} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.479449] env[61824]: INFO nova.virt.vmwareapi.ds_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk. [ 1016.480205] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2c9bc0-2e50-4746-b46d-121fe5226939 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.504724] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.505165] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd02d02e-44a9-4ae0-a534-749a039b0c3b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.522547] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1016.522547] env[61824]: value = "task-1276021" [ 1016.522547] env[61824]: _type = "Task" [ 1016.522547] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.529610] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276021, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.616886] env[61824]: DEBUG nova.compute.utils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.618384] env[61824]: DEBUG nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.618570] env[61824]: DEBUG nova.network.neutron [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1016.657345] env[61824]: DEBUG nova.policy [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8763f64818e24fad80ce232ac7d4463f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '629276280a234e7a8b810e6e74779fff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1016.887190] env[61824]: DEBUG nova.network.neutron [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Successfully created port: b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.032645] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276021, 'name': ReconfigVM_Task, 'duration_secs': 0.300889} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.033391] env[61824]: DEBUG nova.network.neutron [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.035469] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.036018] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c22748-cad3-4071-a3e0-7f0a2b715172 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.060346] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85b4654e-66c5-4ba6-999d-6f3adeea22c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.078092] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1017.078092] env[61824]: value = "task-1276022" [ 1017.078092] env[61824]: _type = "Task" [ 1017.078092] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.088090] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276022, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.121817] env[61824]: DEBUG nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.267157] env[61824]: DEBUG nova.network.neutron [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Updating instance_info_cache with network_info: [{"id": "d998ac38-5777-435b-96bc-d96f4bbcb6b4", "address": "fa:16:3e:eb:9c:b7", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd998ac38-57", "ovs_interfaceid": "d998ac38-5777-435b-96bc-d96f4bbcb6b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.587313] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276022, 'name': ReconfigVM_Task, 'duration_secs': 0.152036} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.587606] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.587860] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0671700a-a951-4780-bf76-b13db895b76a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.593606] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1017.593606] env[61824]: value = "task-1276023" [ 1017.593606] env[61824]: _type = "Task" [ 1017.593606] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.601931] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276023, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.770335] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "refresh_cache-d97da7c1-d20d-45db-a03d-c9d4ceb8c235" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.770663] env[61824]: DEBUG nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Instance network_info: |[{"id": "d998ac38-5777-435b-96bc-d96f4bbcb6b4", "address": "fa:16:3e:eb:9c:b7", "network": {"id": "a44bc585-0686-4f35-ac3b-f10fc6cb8977", "bridge": "br-int", "label": "tempest-ServersTestJSON-748032558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e756c6817698464a9da391d4dd364946", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd998ac38-57", "ovs_interfaceid": "d998ac38-5777-435b-96bc-d96f4bbcb6b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1017.771090] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:9c:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd998ac38-5777-435b-96bc-d96f4bbcb6b4', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.778553] env[61824]: DEBUG oslo.service.loopingcall [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.778774] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1017.779019] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f9b720c-2792-4bc2-9113-465667f54bb5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.799249] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.799249] env[61824]: value = "task-1276024" [ 1017.799249] env[61824]: _type = "Task" [ 1017.799249] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.806521] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276024, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.103397] env[61824]: DEBUG oslo_vmware.api [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276023, 'name': PowerOnVM_Task, 'duration_secs': 0.378221} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.103683] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.106534] env[61824]: DEBUG nova.compute.manager [None req-4f85c6ae-ea3e-4133-b9c2-fe3ea8740749 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.107352] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b65d444-bfbe-47d5-bf9e-e4d84990b993 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.131119] env[61824]: DEBUG nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.156197] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.156454] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.156618] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.156803] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.156953] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.157122] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.157382] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.157565] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.157737] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.157904] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.158096] env[61824]: DEBUG nova.virt.hardware [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.159145] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c9451c-e882-483e-9658-30b7af23e38e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.167034] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0530cb99-28e4-4f47-a1aa-bb3504feb33d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.309250] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276024, 'name': CreateVM_Task, 'duration_secs': 0.29574} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.309487] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.310214] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.310431] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.310813] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.311099] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43018985-251f-4baf-b190-34589450ec2a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.315517] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1018.315517] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]525e8398-2d94-20f0-1122-98df1e79fb0b" [ 1018.315517] env[61824]: _type = "Task" [ 1018.315517] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.322792] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525e8398-2d94-20f0-1122-98df1e79fb0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.491053] env[61824]: DEBUG nova.compute.manager [req-e56b4d22-10fb-4906-96ad-772a40f61afd req-06c5faa3-9852-4549-852f-660ecc3dc348 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Received event network-vif-plugged-b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.491217] env[61824]: DEBUG oslo_concurrency.lockutils [req-e56b4d22-10fb-4906-96ad-772a40f61afd req-06c5faa3-9852-4549-852f-660ecc3dc348 service nova] Acquiring lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.491429] env[61824]: DEBUG oslo_concurrency.lockutils [req-e56b4d22-10fb-4906-96ad-772a40f61afd req-06c5faa3-9852-4549-852f-660ecc3dc348 service nova] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.491610] env[61824]: DEBUG oslo_concurrency.lockutils [req-e56b4d22-10fb-4906-96ad-772a40f61afd req-06c5faa3-9852-4549-852f-660ecc3dc348 service nova] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.491783] env[61824]: DEBUG nova.compute.manager [req-e56b4d22-10fb-4906-96ad-772a40f61afd req-06c5faa3-9852-4549-852f-660ecc3dc348 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] No waiting events found dispatching network-vif-plugged-b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.491951] env[61824]: WARNING nova.compute.manager [req-e56b4d22-10fb-4906-96ad-772a40f61afd req-06c5faa3-9852-4549-852f-660ecc3dc348 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Received unexpected event network-vif-plugged-b8687d0f-80a9-409d-895f-553d2a9bdc70 for instance with vm_state building and task_state spawning. [ 1018.624193] env[61824]: DEBUG nova.network.neutron [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Successfully updated port: b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.827156] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]525e8398-2d94-20f0-1122-98df1e79fb0b, 'name': SearchDatastore_Task, 'duration_secs': 0.00851} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.827562] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.827758] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.828017] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.828177] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.828361] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.828628] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-61571623-2cd4-44ea-bf28-04a2cdc58865 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.836565] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.836752] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1018.837485] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9bb1b58-fae2-49bc-9050-7d8ed3e9e03b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.842531] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1018.842531] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52b7d621-b9d8-6d00-6eb2-52cbc7bd0c0d" [ 1018.842531] env[61824]: _type = "Task" [ 1018.842531] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.850106] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b7d621-b9d8-6d00-6eb2-52cbc7bd0c0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.126643] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.126795] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.126950] env[61824]: DEBUG nova.network.neutron [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1019.353555] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b7d621-b9d8-6d00-6eb2-52cbc7bd0c0d, 'name': SearchDatastore_Task, 'duration_secs': 0.008713} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.354367] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0a52eac-71cf-48d3-a1dd-472621d186f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.359897] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1019.359897] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f0ecf6-6c66-485e-f2bf-d549fbf7a6a1" [ 1019.359897] env[61824]: _type = "Task" [ 1019.359897] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.367579] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f0ecf6-6c66-485e-f2bf-d549fbf7a6a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.657583] env[61824]: DEBUG nova.network.neutron [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1019.777860] env[61824]: DEBUG nova.network.neutron [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Updating instance_info_cache with network_info: [{"id": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "address": "fa:16:3e:4f:91:34", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8687d0f-80", "ovs_interfaceid": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.870632] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f0ecf6-6c66-485e-f2bf-d549fbf7a6a1, 'name': SearchDatastore_Task, 'duration_secs': 0.009333} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.870935] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.871312] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] d97da7c1-d20d-45db-a03d-c9d4ceb8c235/d97da7c1-d20d-45db-a03d-c9d4ceb8c235.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.871427] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a542e362-ce4b-4734-9e7a-fd23ce308498 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.878481] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1019.878481] env[61824]: value = "task-1276025" [ 1019.878481] env[61824]: _type = "Task" [ 1019.878481] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.885751] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276025, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.281085] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.281261] env[61824]: DEBUG nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Instance network_info: |[{"id": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "address": "fa:16:3e:4f:91:34", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8687d0f-80", "ovs_interfaceid": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.281607] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:91:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8687d0f-80a9-409d-895f-553d2a9bdc70', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.289204] env[61824]: DEBUG oslo.service.loopingcall [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.289485] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.289732] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-234a1140-815f-4439-b37b-d1c18d96581a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.310586] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.310586] env[61824]: value = "task-1276026" [ 1020.310586] env[61824]: _type = "Task" [ 1020.310586] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.317905] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276026, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.388991] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276025, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.419865} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.389311] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] d97da7c1-d20d-45db-a03d-c9d4ceb8c235/d97da7c1-d20d-45db-a03d-c9d4ceb8c235.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.389562] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.389837] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7752f01c-3035-4e22-b615-4dd38bdb3734 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.395623] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1020.395623] env[61824]: value = "task-1276027" [ 1020.395623] env[61824]: _type = "Task" [ 1020.395623] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.398976] env[61824]: INFO nova.compute.manager [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Rescuing [ 1020.399226] env[61824]: DEBUG oslo_concurrency.lockutils [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.399415] env[61824]: DEBUG oslo_concurrency.lockutils [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.399643] env[61824]: DEBUG nova.network.neutron [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1020.405447] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276027, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.518154] env[61824]: DEBUG nova.compute.manager [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Received event network-changed-b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.518395] env[61824]: DEBUG nova.compute.manager [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Refreshing instance network info cache due to event network-changed-b8687d0f-80a9-409d-895f-553d2a9bdc70. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.518622] env[61824]: DEBUG oslo_concurrency.lockutils [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] Acquiring lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.518771] env[61824]: DEBUG oslo_concurrency.lockutils [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] Acquired lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.518936] env[61824]: DEBUG nova.network.neutron [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Refreshing network info cache for port b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.822126] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276026, 'name': CreateVM_Task, 'duration_secs': 0.350377} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.822288] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.823085] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.823206] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.823492] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1020.823744] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ba9a7cd-892f-44fe-9e4c-ceea5a1cdeec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.828196] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1020.828196] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5247334c-404a-b07c-c26e-333798f3e69f" [ 1020.828196] env[61824]: _type = "Task" [ 1020.828196] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.835310] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5247334c-404a-b07c-c26e-333798f3e69f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.906758] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276027, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059003} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.907378] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1020.908174] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619469d0-474b-4cf2-afb7-da45745107e1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.929725] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] d97da7c1-d20d-45db-a03d-c9d4ceb8c235/d97da7c1-d20d-45db-a03d-c9d4ceb8c235.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.930027] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26627dd6-7817-41e1-ac65-1e385b43d909 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.951538] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1020.951538] env[61824]: value = "task-1276028" [ 1020.951538] env[61824]: _type = "Task" [ 1020.951538] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.959755] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276028, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.292986] env[61824]: DEBUG nova.network.neutron [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Updating instance_info_cache with network_info: [{"id": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "address": "fa:16:3e:90:3d:72", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a20cdd6-6c", "ovs_interfaceid": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.339069] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5247334c-404a-b07c-c26e-333798f3e69f, 'name': SearchDatastore_Task, 'duration_secs': 0.031562} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.339069] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.339281] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.339571] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.339716] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.339895] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.340172] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8904e231-4caf-4bf4-882f-0d676eb82a40 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.353640] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.353820] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.354522] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3c02c07-feaa-46f3-ab9c-6ea3d3daceee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.359650] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1021.359650] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5231f91f-96a7-4d8d-f5b5-d414f7445acd" [ 1021.359650] env[61824]: _type = "Task" [ 1021.359650] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.363013] env[61824]: DEBUG nova.network.neutron [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Updated VIF entry in instance network info cache for port b8687d0f-80a9-409d-895f-553d2a9bdc70. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1021.363347] env[61824]: DEBUG nova.network.neutron [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Updating instance_info_cache with network_info: [{"id": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "address": "fa:16:3e:4f:91:34", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8687d0f-80", "ovs_interfaceid": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.369039] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231f91f-96a7-4d8d-f5b5-d414f7445acd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.461941] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276028, 'name': ReconfigVM_Task, 'duration_secs': 0.262781} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.462307] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Reconfigured VM instance instance-00000065 to attach disk [datastore2] d97da7c1-d20d-45db-a03d-c9d4ceb8c235/d97da7c1-d20d-45db-a03d-c9d4ceb8c235.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.463037] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d852b0e-e555-4007-b46a-ae22eb3e432b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.468789] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1021.468789] env[61824]: value = "task-1276029" [ 1021.468789] env[61824]: _type = "Task" [ 1021.468789] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.476222] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276029, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.796335] env[61824]: DEBUG oslo_concurrency.lockutils [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.865759] env[61824]: DEBUG oslo_concurrency.lockutils [req-10074398-6588-48ee-8e53-ff66909e32d1 req-0e14065e-75d5-4c38-80f4-3852bafdc32f service nova] Releasing lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.869581] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231f91f-96a7-4d8d-f5b5-d414f7445acd, 'name': SearchDatastore_Task, 'duration_secs': 0.082958} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.870325] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96fa75b4-ac95-41ea-aeaa-d92823ee977b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.875242] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1021.875242] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52031684-ed61-9e56-40a2-1a71536aadf8" [ 1021.875242] env[61824]: _type = "Task" [ 1021.875242] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.882305] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52031684-ed61-9e56-40a2-1a71536aadf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.979489] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276029, 'name': Rename_Task, 'duration_secs': 0.131839} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.979892] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1021.979958] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-abb6ce2a-168c-4691-9176-354346d6da28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.985922] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1021.985922] env[61824]: value = "task-1276030" [ 1021.985922] env[61824]: _type = "Task" [ 1021.985922] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.994966] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276030, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.324642] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.324836] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97ccddf2-1a3c-4f15-aa77-6a841f21f1b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.332258] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1022.332258] env[61824]: value = "task-1276031" [ 1022.332258] env[61824]: _type = "Task" [ 1022.332258] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.340759] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.385871] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52031684-ed61-9e56-40a2-1a71536aadf8, 'name': SearchDatastore_Task, 'duration_secs': 0.060502} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.386192] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.386459] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 76303768-1bb1-4a82-8ca8-df301c19d02d/76303768-1bb1-4a82-8ca8-df301c19d02d.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.386740] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7814a5a2-19fc-4906-b814-2eed3908be95 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.393720] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1022.393720] env[61824]: value = "task-1276032" [ 1022.393720] env[61824]: _type = "Task" [ 1022.393720] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.402439] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276032, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.495743] env[61824]: DEBUG oslo_vmware.api [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276030, 'name': PowerOnVM_Task, 'duration_secs': 0.415307} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.495957] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.496177] env[61824]: INFO nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Took 7.97 seconds to spawn the instance on the hypervisor. [ 1022.496359] env[61824]: DEBUG nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.497135] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b3bfd3-cb34-4b9b-8152-4782eda1f586 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.842065] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276031, 'name': PowerOffVM_Task, 'duration_secs': 0.228157} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.842366] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.843146] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84f2c73-e0ea-43db-99c0-fe729be381c2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.861177] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d56454-cddd-4963-8a73-e8bf58468a61 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.887107] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1022.887503] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6725bdc-8afc-47a7-af10-61443afa0531 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.893349] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1022.893349] env[61824]: value = "task-1276033" [ 1022.893349] env[61824]: _type = "Task" [ 1022.893349] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.906912] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276032, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.431097} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.907237] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 76303768-1bb1-4a82-8ca8-df301c19d02d/76303768-1bb1-4a82-8ca8-df301c19d02d.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1022.907484] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1022.907778] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1022.907980] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1022.908247] env[61824]: DEBUG oslo_concurrency.lockutils [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.908418] env[61824]: DEBUG oslo_concurrency.lockutils [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.908618] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1022.908857] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d73ebde4-cf5b-4cee-8283-33614e112580 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.910624] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1555d103-7a6c-47db-b310-b759f24aba6b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.916094] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1022.916094] env[61824]: value = "task-1276034" [ 1022.916094] env[61824]: _type = "Task" [ 1022.916094] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.919912] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1022.920131] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1022.921269] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f1096b5-d1f9-42ab-91d3-91f3e62d09ad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.928567] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276034, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.931702] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1022.931702] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5281f1f6-45f9-b3bb-4e8e-26e9dbfbda34" [ 1022.931702] env[61824]: _type = "Task" [ 1022.931702] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.939029] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5281f1f6-45f9-b3bb-4e8e-26e9dbfbda34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.012753] env[61824]: INFO nova.compute.manager [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Took 14.22 seconds to build instance. [ 1023.425746] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276034, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05995} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.426233] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.427115] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13be0518-4508-4cd2-9a7e-6bc88bfba4d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.449628] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 76303768-1bb1-4a82-8ca8-df301c19d02d/76303768-1bb1-4a82-8ca8-df301c19d02d.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.452993] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cbe3adc-6d9e-497a-80e2-40920ea1d86c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.473717] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5281f1f6-45f9-b3bb-4e8e-26e9dbfbda34, 'name': SearchDatastore_Task, 'duration_secs': 0.00836} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.476372] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1023.476372] env[61824]: value = "task-1276035" [ 1023.476372] env[61824]: _type = "Task" [ 1023.476372] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.476372] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-619d15a6-f93a-47e9-955e-668a4b8a77b9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.486619] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276035, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.487519] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1023.487519] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52426c75-5e92-871a-28c4-3392367ac06b" [ 1023.487519] env[61824]: _type = "Task" [ 1023.487519] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.495178] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52426c75-5e92-871a-28c4-3392367ac06b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.504876] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.514574] env[61824]: DEBUG oslo_concurrency.lockutils [None req-fc641054-8ee8-43a3-bd25-baeca122afa4 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.727s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.515015] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.010s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.515338] env[61824]: DEBUG nova.compute.manager [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.518016] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300bd747-939b-4a26-bce3-8c2c72064094 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.523129] env[61824]: DEBUG nova.compute.manager [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1023.523690] env[61824]: DEBUG nova.objects.instance [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'flavor' on Instance uuid d97da7c1-d20d-45db-a03d-c9d4ceb8c235 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.988608] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276035, 'name': ReconfigVM_Task, 'duration_secs': 0.274553} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.991610] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 76303768-1bb1-4a82-8ca8-df301c19d02d/76303768-1bb1-4a82-8ca8-df301c19d02d.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.992354] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e467dc92-23a0-423b-84c8-06f751b63eca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.999422] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52426c75-5e92-871a-28c4-3392367ac06b, 'name': SearchDatastore_Task, 'duration_secs': 0.024265} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.001030] env[61824]: DEBUG oslo_concurrency.lockutils [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.001030] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk. {{(pid=61824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1024.001251] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1024.001251] env[61824]: value = "task-1276036" [ 1024.001251] env[61824]: _type = "Task" [ 1024.001251] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.001346] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-15c212d1-5a52-4cb1-9181-f67926796f5c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.010734] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276036, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.011835] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1024.011835] env[61824]: value = "task-1276037" [ 1024.011835] env[61824]: _type = "Task" [ 1024.011835] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.019117] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.029425] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.029425] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a2f0815-5768-4e34-b7d5-9ffa42f6657c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.034700] env[61824]: DEBUG oslo_vmware.api [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1024.034700] env[61824]: value = "task-1276038" [ 1024.034700] env[61824]: _type = "Task" [ 1024.034700] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.043878] env[61824]: DEBUG oslo_vmware.api [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.514716] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276036, 'name': Rename_Task, 'duration_secs': 0.140954} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.518138] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1024.518435] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dff51388-e428-4c1b-982b-17bfd461d21a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.526199] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505895} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.527445] env[61824]: INFO nova.virt.vmwareapi.ds_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk. [ 1024.527877] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1024.527877] env[61824]: value = "task-1276039" [ 1024.527877] env[61824]: _type = "Task" [ 1024.527877] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.528512] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff3df18-3e4c-4a80-b1f9-1f5973206038 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.540540] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.560711] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.561423] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22405a8c-77ec-4868-a380-2e4dda97f591 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.577528] env[61824]: DEBUG oslo_vmware.api [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276038, 'name': PowerOffVM_Task, 'duration_secs': 0.186181} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.578187] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1024.579045] env[61824]: DEBUG nova.compute.manager [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.579122] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77266237-9ef4-4c05-8de8-7ae7f3e3c28f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.585686] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1024.585686] env[61824]: value = "task-1276040" [ 1024.585686] env[61824]: _type = "Task" [ 1024.585686] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.586475] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "a44b2d44-ecab-412a-b22e-c5a720b59631" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.586694] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.598504] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276040, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.040919] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276039, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.096373] env[61824]: DEBUG nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1025.101076] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276040, 'name': ReconfigVM_Task, 'duration_secs': 0.349282} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.101076] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9732e33a-d128-4a0f-9a49-457e2362ef9c tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.586s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.101806] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.102781] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707a47eb-de52-424f-af93-c4d94fea0af3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.131651] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f023b4a1-0c8c-4878-8c92-7cfa0e57cf14 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.146572] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1025.146572] env[61824]: value = "task-1276041" [ 1025.146572] env[61824]: _type = "Task" [ 1025.146572] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.155079] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276041, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.541725] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276039, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.622253] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.622546] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.624068] env[61824]: INFO nova.compute.claims [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1025.656754] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276041, 'name': ReconfigVM_Task, 'duration_secs': 0.272038} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.657074] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.657369] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ab440e1-0e55-4a5e-bf3f-650796d4e7af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.663640] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1025.663640] env[61824]: value = "task-1276042" [ 1025.663640] env[61824]: _type = "Task" [ 1025.663640] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.672166] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.041921] env[61824]: DEBUG oslo_vmware.api [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276039, 'name': PowerOnVM_Task, 'duration_secs': 1.100535} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.042648] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.042648] env[61824]: INFO nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Took 7.91 seconds to spawn the instance on the hypervisor. [ 1026.042648] env[61824]: DEBUG nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.043400] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0c6b42-c6a0-4c10-817c-aeb31d3ccd37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.173247] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276042, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.283255] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.283605] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.283783] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.284195] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.284195] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.286382] env[61824]: INFO nova.compute.manager [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Terminating instance [ 1026.288225] env[61824]: DEBUG nova.compute.manager [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.288416] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.289253] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a095fb3e-c1dd-4b31-9c66-dc070088920d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.296660] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1026.296882] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da02e569-bca7-48bd-815b-a771197bd5eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.370596] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.370826] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.371027] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleting the datastore file [datastore2] d97da7c1-d20d-45db-a03d-c9d4ceb8c235 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.371287] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33038ebc-9c59-4891-be61-d363f206d0a0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.376965] env[61824]: DEBUG oslo_vmware.api [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1026.376965] env[61824]: value = "task-1276044" [ 1026.376965] env[61824]: _type = "Task" [ 1026.376965] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.384205] env[61824]: DEBUG oslo_vmware.api [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276044, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.561849] env[61824]: INFO nova.compute.manager [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Took 12.63 seconds to build instance. [ 1026.673438] env[61824]: DEBUG oslo_vmware.api [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276042, 'name': PowerOnVM_Task, 'duration_secs': 0.617528} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.673747] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.676394] env[61824]: DEBUG nova.compute.manager [None req-058112fc-317d-4485-a09f-11e04d8d5504 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.677186] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e75eaa-f0eb-490a-bc94-c4d12939a0c5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.750161] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46fe8d7-6f64-427b-8b94-83992515e736 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.758077] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179595b9-b5a5-4794-baf3-138eca395971 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.787862] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13109ba-74d8-471e-9309-d494ed62c089 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.795423] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2a4c52-0fdd-4384-9848-490518b8b81a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.808808] env[61824]: DEBUG nova.compute.provider_tree [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1026.887226] env[61824]: DEBUG oslo_vmware.api [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276044, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.063389] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a9d50128-c016-46ee-9861-563b91f4aa51 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.143s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.312051] env[61824]: DEBUG nova.scheduler.client.report [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.388366] env[61824]: DEBUG oslo_vmware.api [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276044, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.562465} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.388630] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.388850] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.389048] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.389242] env[61824]: INFO nova.compute.manager [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1027.389535] env[61824]: DEBUG oslo.service.loopingcall [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.389739] env[61824]: DEBUG nova.compute.manager [-] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.389844] env[61824]: DEBUG nova.network.neutron [-] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1027.565172] env[61824]: INFO nova.compute.manager [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Unrescuing [ 1027.565479] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.565638] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.565808] env[61824]: DEBUG nova.network.neutron [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1027.630776] env[61824]: DEBUG nova.compute.manager [req-5233f6d5-814c-48f0-ba82-968e6cff862c req-a8aeb3dc-e825-4e05-a34a-221983a720d7 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Received event network-vif-deleted-d998ac38-5777-435b-96bc-d96f4bbcb6b4 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.630906] env[61824]: INFO nova.compute.manager [req-5233f6d5-814c-48f0-ba82-968e6cff862c req-a8aeb3dc-e825-4e05-a34a-221983a720d7 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Neutron deleted interface d998ac38-5777-435b-96bc-d96f4bbcb6b4; detaching it from the instance and deleting it from the info cache [ 1027.630996] env[61824]: DEBUG nova.network.neutron [req-5233f6d5-814c-48f0-ba82-968e6cff862c req-a8aeb3dc-e825-4e05-a34a-221983a720d7 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.667761] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "76303768-1bb1-4a82-8ca8-df301c19d02d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.668044] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.668233] env[61824]: INFO nova.compute.manager [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Shelving [ 1027.816886] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.817516] env[61824]: DEBUG nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1028.109986] env[61824]: DEBUG nova.network.neutron [-] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.134173] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e81d4683-f12a-49b1-83f1-cd85be572841 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.145515] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7896f0-c1f1-40d9-ab9e-8eb4c352713b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.181224] env[61824]: DEBUG nova.compute.manager [req-5233f6d5-814c-48f0-ba82-968e6cff862c req-a8aeb3dc-e825-4e05-a34a-221983a720d7 service nova] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Detach interface failed, port_id=d998ac38-5777-435b-96bc-d96f4bbcb6b4, reason: Instance d97da7c1-d20d-45db-a03d-c9d4ceb8c235 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1028.186888] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1028.187620] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.187840] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cada5023-733d-4d79-bf0b-64ea71637513 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.189430] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1028.189657] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1028.199021] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1028.199021] env[61824]: value = "task-1276045" [ 1028.199021] env[61824]: _type = "Task" [ 1028.199021] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.209195] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.314053] env[61824]: DEBUG nova.network.neutron [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Updating instance_info_cache with network_info: [{"id": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "address": "fa:16:3e:90:3d:72", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a20cdd6-6c", "ovs_interfaceid": "2a20cdd6-6c0a-46d2-b5a7-1b78e2476561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.322191] env[61824]: DEBUG nova.compute.utils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1028.323599] env[61824]: DEBUG nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1028.323770] env[61824]: DEBUG nova.network.neutron [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1028.376217] env[61824]: DEBUG nova.policy [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad4e8224518048e583d27b94721b3c22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd21256931aa54010b72beca4fb798f63', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1028.614341] env[61824]: INFO nova.compute.manager [-] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Took 1.22 seconds to deallocate network for instance. [ 1028.709025] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276045, 'name': PowerOffVM_Task, 'duration_secs': 0.2168} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.709290] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1028.710255] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614daf8b-cc72-4a23-a310-05483582c7d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.731344] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0c7668-dcf9-422c-b148-585bee359868 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.756479] env[61824]: DEBUG nova.network.neutron [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Successfully created port: 578e7e39-b803-41c6-890e-a1867602c4b9 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1028.817026] env[61824]: DEBUG oslo_concurrency.lockutils [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-4700de90-7230-43a5-87ae-f2c7ed9d4ea6" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.817702] env[61824]: DEBUG nova.objects.instance [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lazy-loading 'flavor' on Instance uuid 4700de90-7230-43a5-87ae-f2c7ed9d4ea6 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.828074] env[61824]: DEBUG nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1029.120841] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.121180] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.121307] env[61824]: DEBUG nova.objects.instance [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'resources' on Instance uuid d97da7c1-d20d-45db-a03d-c9d4ceb8c235 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.240762] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Creating Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1029.241094] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ef6ddada-4013-4426-b0a6-fc35144f956d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.250722] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1029.250722] env[61824]: value = "task-1276046" [ 1029.250722] env[61824]: _type = "Task" [ 1029.250722] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.259409] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276046, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.323576] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10063395-9b49-4cb7-9091-539a4dd7e28e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.351642] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.352013] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cb80c44-a7a8-4b97-9690-fa748a1fbefd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.359790] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1029.359790] env[61824]: value = "task-1276047" [ 1029.359790] env[61824]: _type = "Task" [ 1029.359790] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.368487] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276047, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.742017] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cbf007-f9d3-4301-91a8-96a3fb9c315f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.750406] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1088cbf6-0bda-4635-997b-2590cacd1d12 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.761915] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276046, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.787119] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db196c9-455e-4e7b-9e18-14bf460dcd81 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.795645] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b40f65-6a28-4df6-975d-82a8deabf1ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.810245] env[61824]: DEBUG nova.compute.provider_tree [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.853852] env[61824]: DEBUG nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1029.871828] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276047, 'name': PowerOffVM_Task, 'duration_secs': 0.255671} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.872235] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.877582] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1029.879903] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb119ba9-ed9d-4782-9618-afa8e984d396 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.900486] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1029.900789] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1029.900972] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1029.901249] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1029.901353] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1029.901483] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1029.901708] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1029.901922] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1029.902141] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1029.902312] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1029.902502] env[61824]: DEBUG nova.virt.hardware [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1029.903435] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a638c9d2-1e89-4be4-a077-4528d4ec0384 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.908075] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1029.908075] env[61824]: value = "task-1276048" [ 1029.908075] env[61824]: _type = "Task" [ 1029.908075] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.915728] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e1011c-fe3b-42f9-9fa4-c4b3dadb5adf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.923265] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276048, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.184594] env[61824]: DEBUG nova.compute.manager [req-d8a7d112-34ad-4b1d-9f3b-3268b0428b23 req-006b7d3b-5981-4d35-b4cd-65d37ba6e6f0 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Received event network-vif-plugged-578e7e39-b803-41c6-890e-a1867602c4b9 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.184954] env[61824]: DEBUG oslo_concurrency.lockutils [req-d8a7d112-34ad-4b1d-9f3b-3268b0428b23 req-006b7d3b-5981-4d35-b4cd-65d37ba6e6f0 service nova] Acquiring lock "a44b2d44-ecab-412a-b22e-c5a720b59631-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.185142] env[61824]: DEBUG oslo_concurrency.lockutils [req-d8a7d112-34ad-4b1d-9f3b-3268b0428b23 req-006b7d3b-5981-4d35-b4cd-65d37ba6e6f0 service nova] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.185339] env[61824]: DEBUG oslo_concurrency.lockutils [req-d8a7d112-34ad-4b1d-9f3b-3268b0428b23 req-006b7d3b-5981-4d35-b4cd-65d37ba6e6f0 service nova] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.185521] env[61824]: DEBUG nova.compute.manager [req-d8a7d112-34ad-4b1d-9f3b-3268b0428b23 req-006b7d3b-5981-4d35-b4cd-65d37ba6e6f0 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] No waiting events found dispatching network-vif-plugged-578e7e39-b803-41c6-890e-a1867602c4b9 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.185697] env[61824]: WARNING nova.compute.manager [req-d8a7d112-34ad-4b1d-9f3b-3268b0428b23 req-006b7d3b-5981-4d35-b4cd-65d37ba6e6f0 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Received unexpected event network-vif-plugged-578e7e39-b803-41c6-890e-a1867602c4b9 for instance with vm_state building and task_state spawning. [ 1030.200569] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Didn't find any instances for network info cache update. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1030.200772] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1030.200937] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1030.201461] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1030.204914] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1030.204914] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1030.204914] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1030.204914] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1030.204914] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1030.264874] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276046, 'name': CreateSnapshot_Task, 'duration_secs': 0.624349} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.265232] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Created Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1030.266157] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3405e825-233d-4cdb-9e2f-f3721d166e26 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.286310] env[61824]: DEBUG nova.network.neutron [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Successfully updated port: 578e7e39-b803-41c6-890e-a1867602c4b9 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1030.314461] env[61824]: DEBUG nova.scheduler.client.report [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.418921] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276048, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.705919] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.788949] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Creating linked-clone VM from snapshot {{(pid=61824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1030.788949] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-458b4d1a-f843-45a8-a11d-8fee65577ca6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.790664] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-a44b2d44-ecab-412a-b22e-c5a720b59631" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.790877] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-a44b2d44-ecab-412a-b22e-c5a720b59631" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.791113] env[61824]: DEBUG nova.network.neutron [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.801566] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1030.801566] env[61824]: value = "task-1276049" [ 1030.801566] env[61824]: _type = "Task" [ 1030.801566] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.813381] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276049, 'name': CloneVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.819683] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.822221] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.116s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.822504] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.822749] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1030.824066] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1817bf5-ee72-48a4-86a0-1c25c246045e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.835386] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647e1991-d519-4394-ba86-5c32cbba8af7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.844961] env[61824]: INFO nova.scheduler.client.report [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted allocations for instance d97da7c1-d20d-45db-a03d-c9d4ceb8c235 [ 1030.862033] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f65a67b-245d-434e-9c9a-2b582be2e5ef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.868068] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6011ab73-56d4-41ef-bb94-96cc5559fd35 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.902482] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180896MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1030.902690] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.902899] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.918924] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276048, 'name': ReconfigVM_Task, 'duration_secs': 0.817043} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.919254] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1030.919453] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1030.919762] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c3b9919-3fa5-4491-a9b6-6b1caa947a62 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.928586] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1030.928586] env[61824]: value = "task-1276050" [ 1030.928586] env[61824]: _type = "Task" [ 1030.928586] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.940841] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276050, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.315176] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276049, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.338247] env[61824]: DEBUG nova.network.neutron [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1031.365297] env[61824]: DEBUG oslo_concurrency.lockutils [None req-aa783e13-cc01-483d-b5c4-9888d29fd825 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "d97da7c1-d20d-45db-a03d-c9d4ceb8c235" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.082s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.438462] env[61824]: DEBUG oslo_vmware.api [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276050, 'name': PowerOnVM_Task, 'duration_secs': 0.432829} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.438755] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1031.438994] env[61824]: DEBUG nova.compute.manager [None req-f517277b-4ba3-48c5-b4b8-41083895f344 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1031.439837] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fdd603-110f-42c8-a2ac-dc4cfa2732f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.487283] env[61824]: DEBUG nova.network.neutron [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Updating instance_info_cache with network_info: [{"id": "578e7e39-b803-41c6-890e-a1867602c4b9", "address": "fa:16:3e:f6:a8:b9", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap578e7e39-b8", "ovs_interfaceid": "578e7e39-b803-41c6-890e-a1867602c4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.813761] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276049, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.935007] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1031.935007] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8aa00861-bca6-4861-a5e5-0c538155c9ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1031.935007] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 5bf9f3b5-7084-4c5a-952b-d90c337ad10a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1031.935007] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 4700de90-7230-43a5-87ae-f2c7ed9d4ea6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1031.935007] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 1ddb6d7a-d8e1-492d-84c9-d61e27faa105 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1031.935007] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 76303768-1bb1-4a82-8ca8-df301c19d02d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1031.935406] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance a44b2d44-ecab-412a-b22e-c5a720b59631 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1031.935406] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1031.935503] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1031.989913] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-a44b2d44-ecab-412a-b22e-c5a720b59631" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.990479] env[61824]: DEBUG nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Instance network_info: |[{"id": "578e7e39-b803-41c6-890e-a1867602c4b9", "address": "fa:16:3e:f6:a8:b9", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap578e7e39-b8", "ovs_interfaceid": "578e7e39-b803-41c6-890e-a1867602c4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1031.991040] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:a8:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '578e7e39-b803-41c6-890e-a1867602c4b9', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1032.000617] env[61824]: DEBUG oslo.service.loopingcall [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.003115] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1032.003856] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cce9e436-199b-4a23-bb4e-e0c9c383cafc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.027609] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1032.027609] env[61824]: value = "task-1276051" [ 1032.027609] env[61824]: _type = "Task" [ 1032.027609] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.038477] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276051, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.055025] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffcd008-9fe8-4560-a228-33b072616bfe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.063826] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f069d544-8a9a-4a8d-a74a-826a53d7564a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.095590] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.095930] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.096216] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.096428] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.096608] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.099045] env[61824]: INFO nova.compute.manager [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Terminating instance [ 1032.101570] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6ceb53-9124-48ae-bae4-073bd46f5c84 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.104866] env[61824]: DEBUG nova.compute.manager [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.105151] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.106486] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81fa7ff1-f0fb-484a-9f96-2c5393e825d7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.118095] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf639034-0c43-4378-b29c-1554f7ebae88 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.122369] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.122667] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79ffb2d3-c6d7-4763-b54b-d1e714fc59c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.137363] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.140459] env[61824]: DEBUG oslo_vmware.api [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1032.140459] env[61824]: value = "task-1276052" [ 1032.140459] env[61824]: _type = "Task" [ 1032.140459] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.150778] env[61824]: DEBUG oslo_vmware.api [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.220905] env[61824]: DEBUG nova.compute.manager [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Received event network-changed-578e7e39-b803-41c6-890e-a1867602c4b9 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.221204] env[61824]: DEBUG nova.compute.manager [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Refreshing instance network info cache due to event network-changed-578e7e39-b803-41c6-890e-a1867602c4b9. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.221352] env[61824]: DEBUG oslo_concurrency.lockutils [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] Acquiring lock "refresh_cache-a44b2d44-ecab-412a-b22e-c5a720b59631" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.221502] env[61824]: DEBUG oslo_concurrency.lockutils [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] Acquired lock "refresh_cache-a44b2d44-ecab-412a-b22e-c5a720b59631" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.221703] env[61824]: DEBUG nova.network.neutron [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Refreshing network info cache for port 578e7e39-b803-41c6-890e-a1867602c4b9 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.315487] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276049, 'name': CloneVM_Task, 'duration_secs': 1.464129} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.315928] env[61824]: INFO nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Created linked-clone VM from snapshot [ 1032.316604] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8feb5c48-66cf-4ae6-a04a-0f74bf5b7db3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.327043] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Uploading image ca256eda-b764-491d-b426-e7bdfcf3ba90 {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1032.358879] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1032.358879] env[61824]: value = "vm-274230" [ 1032.358879] env[61824]: _type = "VirtualMachine" [ 1032.358879] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1032.359205] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f1d7d29b-256f-46d0-ac4d-2f4c04dde8cd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.367202] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lease: (returnval){ [ 1032.367202] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e4b6c0-0745-2451-118e-37893deb52d9" [ 1032.367202] env[61824]: _type = "HttpNfcLease" [ 1032.367202] env[61824]: } obtained for exporting VM: (result){ [ 1032.367202] env[61824]: value = "vm-274230" [ 1032.367202] env[61824]: _type = "VirtualMachine" [ 1032.367202] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1032.367542] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the lease: (returnval){ [ 1032.367542] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e4b6c0-0745-2451-118e-37893deb52d9" [ 1032.367542] env[61824]: _type = "HttpNfcLease" [ 1032.367542] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1032.375209] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1032.375209] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e4b6c0-0745-2451-118e-37893deb52d9" [ 1032.375209] env[61824]: _type = "HttpNfcLease" [ 1032.375209] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1032.537734] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276051, 'name': CreateVM_Task, 'duration_secs': 0.354259} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.537877] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1032.538589] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.538762] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.539111] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1032.539369] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13c02ea4-aea6-48f4-a2c6-aa3618d6e09a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.544427] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1032.544427] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f2c67d-499a-5c30-419f-d51171f10d83" [ 1032.544427] env[61824]: _type = "Task" [ 1032.544427] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.551886] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f2c67d-499a-5c30-419f-d51171f10d83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.642264] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.654536] env[61824]: DEBUG oslo_vmware.api [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276052, 'name': PowerOffVM_Task, 'duration_secs': 0.23925} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.655148] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.655148] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1032.655268] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9123e2aa-de47-46d7-ae66-e7948c54440b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.745176] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1032.745522] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1032.745806] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleting the datastore file [datastore2] 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.746167] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-510ebe09-d9a2-4307-b125-a7a8d0751eca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.754108] env[61824]: DEBUG oslo_vmware.api [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for the task: (returnval){ [ 1032.754108] env[61824]: value = "task-1276055" [ 1032.754108] env[61824]: _type = "Task" [ 1032.754108] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.762311] env[61824]: DEBUG oslo_vmware.api [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276055, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.875884] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1032.875884] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e4b6c0-0745-2451-118e-37893deb52d9" [ 1032.875884] env[61824]: _type = "HttpNfcLease" [ 1032.875884] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1032.876217] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1032.876217] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52e4b6c0-0745-2451-118e-37893deb52d9" [ 1032.876217] env[61824]: _type = "HttpNfcLease" [ 1032.876217] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1032.876924] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7baae24c-b3b1-487b-a678-29ed93b402c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.888246] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528d0c3c-3862-f22e-2d93-cfde5725e03d/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1032.888246] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528d0c3c-3862-f22e-2d93-cfde5725e03d/disk-0.vmdk for reading. {{(pid=61824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1032.988674] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a03177a0-73d7-4b8d-b116-6374157281e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.056311] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f2c67d-499a-5c30-419f-d51171f10d83, 'name': SearchDatastore_Task, 'duration_secs': 0.026491} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.056633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.056871] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1033.057125] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.057280] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.057468] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1033.057751] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe317b96-703d-4019-ba88-ad547b03636a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.066405] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1033.066597] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1033.067339] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7763431-2acb-4070-8596-0db3e1cdd450 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.072787] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1033.072787] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52689ae3-1566-a894-08b9-6e5f8723d7ca" [ 1033.072787] env[61824]: _type = "Task" [ 1033.072787] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.082203] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52689ae3-1566-a894-08b9-6e5f8723d7ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.090638] env[61824]: DEBUG nova.network.neutron [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Updated VIF entry in instance network info cache for port 578e7e39-b803-41c6-890e-a1867602c4b9. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1033.091011] env[61824]: DEBUG nova.network.neutron [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Updating instance_info_cache with network_info: [{"id": "578e7e39-b803-41c6-890e-a1867602c4b9", "address": "fa:16:3e:f6:a8:b9", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap578e7e39-b8", "ovs_interfaceid": "578e7e39-b803-41c6-890e-a1867602c4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.150382] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1033.150598] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.248s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.264075] env[61824]: DEBUG oslo_vmware.api [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Task: {'id': task-1276055, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.328205} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.264303] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.264495] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.264721] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.264908] env[61824]: INFO nova.compute.manager [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1033.265176] env[61824]: DEBUG oslo.service.loopingcall [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.265382] env[61824]: DEBUG nova.compute.manager [-] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.265473] env[61824]: DEBUG nova.network.neutron [-] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1033.585506] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52689ae3-1566-a894-08b9-6e5f8723d7ca, 'name': SearchDatastore_Task, 'duration_secs': 0.012237} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.587041] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1eb581c-beb9-4ba4-9c85-3b1a1c2ee54f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.594656] env[61824]: DEBUG oslo_concurrency.lockutils [req-f89ff4d1-fbc1-4799-899e-033df6e49112 req-98011ef9-731a-485b-8c90-15911636b182 service nova] Releasing lock "refresh_cache-a44b2d44-ecab-412a-b22e-c5a720b59631" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.594656] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1033.594656] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f9687c-a64c-356e-2b84-f0100cf590c5" [ 1033.594656] env[61824]: _type = "Task" [ 1033.594656] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.604710] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f9687c-a64c-356e-2b84-f0100cf590c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.039811] env[61824]: DEBUG nova.network.neutron [-] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.106614] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f9687c-a64c-356e-2b84-f0100cf590c5, 'name': SearchDatastore_Task, 'duration_secs': 0.011942} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.106960] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.107303] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1034.107635] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7bbdef15-22c9-415b-b2f8-6df3df3fd228 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.116958] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1034.116958] env[61824]: value = "task-1276056" [ 1034.116958] env[61824]: _type = "Task" [ 1034.116958] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.126775] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276056, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.247845] env[61824]: DEBUG nova.compute.manager [req-ec9acd66-f99c-4d87-b5e2-690a1d9e23a6 req-a3f92fa8-c469-430a-923b-b4a35c22470e service nova] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Received event network-vif-deleted-3e5959a6-69ca-4b0a-b953-d6e291034bb2 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.542912] env[61824]: INFO nova.compute.manager [-] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Took 1.28 seconds to deallocate network for instance. [ 1034.631588] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276056, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.051262] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.051589] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.051840] env[61824]: DEBUG nova.objects.instance [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lazy-loading 'resources' on Instance uuid 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.130470] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276056, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54708} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.130770] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.130986] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.131265] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3e16f9d-2e7b-4307-965b-b6ce3ca79215 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.138870] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1035.138870] env[61824]: value = "task-1276057" [ 1035.138870] env[61824]: _type = "Task" [ 1035.138870] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.148952] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276057, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.286276] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "b8db876d-e430-4c5f-a439-0601dbf4598d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.286520] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.649437] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276057, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107213} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.649437] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1035.650566] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62952e78-4dfb-4b8b-b079-9da21161319c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.672290] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.675123] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f0067ad-0052-467b-95c1-2e4f72fe930e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.694978] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1035.694978] env[61824]: value = "task-1276058" [ 1035.694978] env[61824]: _type = "Task" [ 1035.694978] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.700578] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ab53d5-71df-4a16-8861-9b17b42117c4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.706054] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276058, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.710517] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe9890b-b6a0-4fec-9875-89bdd0fd6f94 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.743014] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f984bf8-0a77-4751-ad8a-55e484905d3a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.751082] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cea85b-6e55-4434-b519-36083b8e1d05 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.764654] env[61824]: DEBUG nova.compute.provider_tree [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.789350] env[61824]: DEBUG nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1036.205233] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276058, 'name': ReconfigVM_Task, 'duration_secs': 0.263563} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.205614] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Reconfigured VM instance instance-00000067 to attach disk [datastore1] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1036.206338] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-348b5014-8412-44fe-b76b-c9154467b479 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.213635] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1036.213635] env[61824]: value = "task-1276059" [ 1036.213635] env[61824]: _type = "Task" [ 1036.213635] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.222074] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276059, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.268489] env[61824]: DEBUG nova.scheduler.client.report [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.308769] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.723686] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276059, 'name': Rename_Task, 'duration_secs': 0.13785} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.724404] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1036.724786] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-232c4199-1b9e-464f-ac75-3bf8db751c16 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.734307] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1036.734307] env[61824]: value = "task-1276060" [ 1036.734307] env[61824]: _type = "Task" [ 1036.734307] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.741199] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.774407] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.778241] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.468s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.780075] env[61824]: INFO nova.compute.claims [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.800762] env[61824]: INFO nova.scheduler.client.report [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Deleted allocations for instance 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576 [ 1037.242702] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276060, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.308319] env[61824]: DEBUG oslo_concurrency.lockutils [None req-248262e6-b8b8-4eae-80dc-4e0b5b0db030 tempest-ServersTestJSON-431435274 tempest-ServersTestJSON-431435274-project-member] Lock "1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.212s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.392213] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.392489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.392672] env[61824]: INFO nova.compute.manager [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Shelving [ 1037.742797] env[61824]: DEBUG oslo_vmware.api [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276060, 'name': PowerOnVM_Task, 'duration_secs': 0.603111} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.743137] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1037.743341] env[61824]: INFO nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Took 7.89 seconds to spawn the instance on the hypervisor. [ 1037.743531] env[61824]: DEBUG nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1037.744364] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afaebff5-86c9-47ea-af9e-85dccf8c4fd6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.899660] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.899947] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2a671e7-0fd5-4113-97fc-d255d3e190d8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.903605] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b5a7e9-1a75-42ed-9b3a-f1f14b106612 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.912423] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dffa768-3d8a-422e-9653-58a2cc8f9f09 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.915996] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1037.915996] env[61824]: value = "task-1276061" [ 1037.915996] env[61824]: _type = "Task" [ 1037.915996] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.948831] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63d3a67-af78-45de-b00a-5a71c36bc269 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.954468] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.959146] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6acf50-eb21-4192-9446-c0dd07763b9a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.972409] env[61824]: DEBUG nova.compute.provider_tree [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.267093] env[61824]: INFO nova.compute.manager [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Took 12.66 seconds to build instance. [ 1038.427376] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276061, 'name': PowerOffVM_Task, 'duration_secs': 0.208304} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.427615] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1038.428428] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-531ca225-e41a-4e7e-892e-7b344eaa500c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.446941] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec69890e-7432-4951-b99b-251159165eb0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.475988] env[61824]: DEBUG nova.scheduler.client.report [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1038.770044] env[61824]: DEBUG oslo_concurrency.lockutils [None req-355f7301-d8c9-44c3-b4f0-65f4d64b6821 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.183s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.960942] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Creating Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1038.961371] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8fdb68d9-df74-4a60-b69f-c6ef3eaa6ceb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.969772] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1038.969772] env[61824]: value = "task-1276062" [ 1038.969772] env[61824]: _type = "Task" [ 1038.969772] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.978739] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276062, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.981112] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.983353] env[61824]: DEBUG nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1039.324682] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "a44b2d44-ecab-412a-b22e-c5a720b59631" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.324907] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.325126] env[61824]: DEBUG nova.compute.manager [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.326123] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e67546-1e46-4c6c-9447-e5cd012f4a45 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.333018] env[61824]: DEBUG nova.compute.manager [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1039.333639] env[61824]: DEBUG nova.objects.instance [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'flavor' on Instance uuid a44b2d44-ecab-412a-b22e-c5a720b59631 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.479462] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276062, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.489191] env[61824]: DEBUG nova.compute.utils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1039.490643] env[61824]: DEBUG nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1039.490809] env[61824]: DEBUG nova.network.neutron [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1039.542527] env[61824]: DEBUG nova.policy [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd64f231608684672aed8882bf15faf7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f9aff6d76e534482b3c4d98034a9591a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1039.836610] env[61824]: DEBUG nova.network.neutron [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Successfully created port: c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.845575] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1039.845965] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8fcdd5a-e862-473c-b41c-8da8ab696495 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.856645] env[61824]: DEBUG oslo_vmware.api [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1039.856645] env[61824]: value = "task-1276063" [ 1039.856645] env[61824]: _type = "Task" [ 1039.856645] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.874499] env[61824]: DEBUG oslo_vmware.api [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276063, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.985039] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276062, 'name': CreateSnapshot_Task, 'duration_secs': 0.978959} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.985608] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Created Snapshot of the VM instance {{(pid=61824) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1039.986810] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819ae2c9-7fb7-4e08-8994-6d9dffcf310a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.995296] env[61824]: DEBUG nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1040.090791] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528d0c3c-3862-f22e-2d93-cfde5725e03d/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1040.091726] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153ce5a1-abc4-4d79-9ba0-460ac0f42418 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.097845] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528d0c3c-3862-f22e-2d93-cfde5725e03d/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1040.100018] env[61824]: ERROR oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528d0c3c-3862-f22e-2d93-cfde5725e03d/disk-0.vmdk due to incomplete transfer. [ 1040.100018] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3728e9ca-3d5b-4987-88fa-d4180c0916a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.105195] env[61824]: DEBUG oslo_vmware.rw_handles [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528d0c3c-3862-f22e-2d93-cfde5725e03d/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1040.105392] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Uploaded image ca256eda-b764-491d-b426-e7bdfcf3ba90 to the Glance image server {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1040.107743] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Destroying the VM {{(pid=61824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1040.107840] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4681a331-bde4-4575-9821-0f267fa1aeae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.113560] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1040.113560] env[61824]: value = "task-1276064" [ 1040.113560] env[61824]: _type = "Task" [ 1040.113560] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.122014] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276064, 'name': Destroy_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.201777] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "358cd7ec-99b4-446c-9e60-2fe50d375893" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.202108] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "358cd7ec-99b4-446c-9e60-2fe50d375893" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.366474] env[61824]: DEBUG oslo_vmware.api [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276063, 'name': PowerOffVM_Task, 'duration_secs': 0.194747} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.366474] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1040.366729] env[61824]: DEBUG nova.compute.manager [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.367376] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a963db8d-ed9c-471a-801b-18c199243ab1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.522108] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Creating linked-clone VM from snapshot {{(pid=61824) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1040.522634] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-86e17fb3-96d1-4588-be0a-2fa95359298e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.530855] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1040.530855] env[61824]: value = "task-1276065" [ 1040.530855] env[61824]: _type = "Task" [ 1040.530855] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.539068] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276065, 'name': CloneVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.623866] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276064, 'name': Destroy_Task, 'duration_secs': 0.292093} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.624198] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Destroyed the VM [ 1040.624368] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Deleting Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1040.624626] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4a6275f2-1be6-4fcb-a800-1c33f0e611ad {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.630614] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1040.630614] env[61824]: value = "task-1276066" [ 1040.630614] env[61824]: _type = "Task" [ 1040.630614] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.638070] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276066, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.704434] env[61824]: DEBUG nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1040.880175] env[61824]: DEBUG oslo_concurrency.lockutils [None req-6f1fa970-b41d-4773-8457-2d57f38e5b99 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.024347] env[61824]: DEBUG nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1041.040807] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276065, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.049576] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.049840] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.050014] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.050196] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.051281] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.051281] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.051281] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.051281] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.051281] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.051281] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.051711] env[61824]: DEBUG nova.virt.hardware [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.052180] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55304887-ccd2-4845-9ac8-0966dada9758 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.061292] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c895c1-2b7f-484c-8c89-3570d63f4fd5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.139849] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276066, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.227628] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.227913] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.229452] env[61824]: INFO nova.compute.claims [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.542698] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276065, 'name': CloneVM_Task} progress is 94%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.640850] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276066, 'name': RemoveSnapshot_Task, 'duration_secs': 0.994157} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.641170] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Deleted Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1041.641453] env[61824]: DEBUG nova.compute.manager [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.642297] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede909b9-d587-4102-aa63-1a087a18797a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.646813] env[61824]: DEBUG nova.compute.manager [req-44692db6-2ae7-4936-8c0e-81f3b8cdbee4 req-1d24d809-0684-4e4a-bde6-cae747bd7c84 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Received event network-vif-plugged-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.647045] env[61824]: DEBUG oslo_concurrency.lockutils [req-44692db6-2ae7-4936-8c0e-81f3b8cdbee4 req-1d24d809-0684-4e4a-bde6-cae747bd7c84 service nova] Acquiring lock "b8db876d-e430-4c5f-a439-0601dbf4598d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.647262] env[61824]: DEBUG oslo_concurrency.lockutils [req-44692db6-2ae7-4936-8c0e-81f3b8cdbee4 req-1d24d809-0684-4e4a-bde6-cae747bd7c84 service nova] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.647433] env[61824]: DEBUG oslo_concurrency.lockutils [req-44692db6-2ae7-4936-8c0e-81f3b8cdbee4 req-1d24d809-0684-4e4a-bde6-cae747bd7c84 service nova] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.647605] env[61824]: DEBUG nova.compute.manager [req-44692db6-2ae7-4936-8c0e-81f3b8cdbee4 req-1d24d809-0684-4e4a-bde6-cae747bd7c84 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] No waiting events found dispatching network-vif-plugged-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1041.647775] env[61824]: WARNING nova.compute.manager [req-44692db6-2ae7-4936-8c0e-81f3b8cdbee4 req-1d24d809-0684-4e4a-bde6-cae747bd7c84 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Received unexpected event network-vif-plugged-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d for instance with vm_state building and task_state spawning. [ 1041.716403] env[61824]: DEBUG nova.network.neutron [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Successfully updated port: c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1042.042492] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276065, 'name': CloneVM_Task, 'duration_secs': 1.464867} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.042848] env[61824]: INFO nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Created linked-clone VM from snapshot [ 1042.043499] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb5253d-04fb-4a9a-9dc6-28a47ab71d34 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.050200] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Uploading image c263e67c-1047-4baa-835b-62b408c6dbe8 {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1042.073146] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1042.073146] env[61824]: value = "vm-274233" [ 1042.073146] env[61824]: _type = "VirtualMachine" [ 1042.073146] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1042.073402] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-91d86a8b-89c9-4076-9c46-4fb1ef05039f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.079895] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease: (returnval){ [ 1042.079895] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52126458-54be-a1be-3d62-97bd8746c52d" [ 1042.079895] env[61824]: _type = "HttpNfcLease" [ 1042.079895] env[61824]: } obtained for exporting VM: (result){ [ 1042.079895] env[61824]: value = "vm-274233" [ 1042.079895] env[61824]: _type = "VirtualMachine" [ 1042.079895] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1042.080163] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the lease: (returnval){ [ 1042.080163] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52126458-54be-a1be-3d62-97bd8746c52d" [ 1042.080163] env[61824]: _type = "HttpNfcLease" [ 1042.080163] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1042.086250] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1042.086250] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52126458-54be-a1be-3d62-97bd8746c52d" [ 1042.086250] env[61824]: _type = "HttpNfcLease" [ 1042.086250] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1042.157563] env[61824]: INFO nova.compute.manager [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Shelve offloading [ 1042.159288] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.159530] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aab59fda-ce5d-4ef7-8748-73bbdaa088e8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.165799] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1042.165799] env[61824]: value = "task-1276068" [ 1042.165799] env[61824]: _type = "Task" [ 1042.165799] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.174503] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.219179] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.219317] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.219464] env[61824]: DEBUG nova.network.neutron [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.261879] env[61824]: INFO nova.compute.manager [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Rebuilding instance [ 1042.293627] env[61824]: DEBUG nova.compute.manager [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.294455] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680b8627-8e9f-4a82-a845-f3ad160de7a9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.334981] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44066d21-ce53-4c82-9a56-c6e240eced5d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.342704] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823774c4-7620-4843-b04d-573502474930 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.371886] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5295c8b3-bf91-4e5e-b181-228fe75cd36c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.379216] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc85d635-f386-4813-8e93-b8a0015ea96f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.392466] env[61824]: DEBUG nova.compute.provider_tree [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.588719] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1042.588719] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52126458-54be-a1be-3d62-97bd8746c52d" [ 1042.588719] env[61824]: _type = "HttpNfcLease" [ 1042.588719] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1042.588991] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1042.588991] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52126458-54be-a1be-3d62-97bd8746c52d" [ 1042.588991] env[61824]: _type = "HttpNfcLease" [ 1042.588991] env[61824]: }. {{(pid=61824) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1042.589747] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafd4fb2-eead-4af7-914d-169d7acd5c1b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.596987] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b914c-fbb0-4f20-10f9-00c1fb456f9f/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1042.597186] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b914c-fbb0-4f20-10f9-00c1fb456f9f/disk-0.vmdk for reading. {{(pid=61824) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1042.676975] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1042.677204] env[61824]: DEBUG nova.compute.manager [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.677960] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b06ae6-1301-4eb7-85b0-929c42dd9e89 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.684956] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c7aa6e2b-bc9a-410b-b860-fe1d1a336a07 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.686519] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.686691] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.686868] env[61824]: DEBUG nova.network.neutron [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1042.773289] env[61824]: DEBUG nova.network.neutron [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1042.807189] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.807492] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c909e5e0-ac35-41f2-a08f-93d86f35c3c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.813727] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1042.813727] env[61824]: value = "task-1276069" [ 1042.813727] env[61824]: _type = "Task" [ 1042.813727] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.822110] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.895296] env[61824]: DEBUG nova.scheduler.client.report [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.214886] env[61824]: DEBUG nova.network.neutron [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.326896] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1043.327216] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1043.328085] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be7e0a3-eadf-47e8-b11b-f73cb67813b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.335186] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1043.335500] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9542cdd1-d845-48fa-aa34-8ca156f81260 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.398821] env[61824]: DEBUG nova.network.neutron [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Updating instance_info_cache with network_info: [{"id": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "address": "fa:16:3e:4f:91:34", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8687d0f-80", "ovs_interfaceid": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.400643] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.401150] env[61824]: DEBUG nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1043.406639] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1043.406639] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1043.406639] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleting the datastore file [datastore1] a44b2d44-ecab-412a-b22e-c5a720b59631 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.406639] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ebcfab8e-0d9e-4de8-9143-4c64fceba1f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.412502] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1043.412502] env[61824]: value = "task-1276071" [ 1043.412502] env[61824]: _type = "Task" [ 1043.412502] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.420767] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276071, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.675638] env[61824]: DEBUG nova.compute.manager [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Received event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.676018] env[61824]: DEBUG nova.compute.manager [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing instance network info cache due to event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1043.676298] env[61824]: DEBUG oslo_concurrency.lockutils [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] Acquiring lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.718930] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.719295] env[61824]: DEBUG nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Instance network_info: |[{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1043.719599] env[61824]: DEBUG oslo_concurrency.lockutils [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] Acquired lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.719872] env[61824]: DEBUG nova.network.neutron [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1043.721205] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:1a:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b399c74-1411-408a-b4cd-84e268ae83fe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c75f0bf1-26d7-46fd-a861-f3ddbf753d2d', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1043.729223] env[61824]: DEBUG oslo.service.loopingcall [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.730281] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1043.730574] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-757d05b3-f7d2-4d22-9deb-ff53e2b17c64 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.750675] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1043.750675] env[61824]: value = "task-1276072" [ 1043.750675] env[61824]: _type = "Task" [ 1043.750675] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.762910] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276072, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.905870] env[61824]: DEBUG nova.compute.utils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.908531] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.911226] env[61824]: DEBUG nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Not allocating networking since 'none' was specified. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1043.925085] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276071, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137006} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.925451] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.925768] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.925966] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.261699] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276072, 'name': CreateVM_Task, 'duration_secs': 0.406226} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.262068] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1044.262742] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.262965] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.263394] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1044.264523] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1044.264844] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cb03bc8-f1d1-4a7d-8b15-ab68af1371bf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.267150] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0b0f1a-d370-4569-af94-1ebdd7ae4bbf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.272638] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1044.272638] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]523080a3-462b-c611-878f-c78c409c458d" [ 1044.272638] env[61824]: _type = "Task" [ 1044.272638] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.275200] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1044.278420] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d119f38-b819-4944-855f-a3a027a33a97 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.285070] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]523080a3-462b-c611-878f-c78c409c458d, 'name': SearchDatastore_Task, 'duration_secs': 0.009066} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.285418] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.285705] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.285959] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.286188] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.286386] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.286705] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b2a9e0f-34a6-4261-a799-3eb9f1923079 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.294076] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.294266] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.295096] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b763b983-ecd9-4dd6-b3a6-41a63844967d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.302089] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1044.302089] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52fb2d8d-fdc0-9eef-4259-253ab64e6484" [ 1044.302089] env[61824]: _type = "Task" [ 1044.302089] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.310806] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fb2d8d-fdc0-9eef-4259-253ab64e6484, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.341831] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1044.342116] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1044.342388] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleting the datastore file [datastore2] 76303768-1bb1-4a82-8ca8-df301c19d02d {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.342778] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c94454d-90df-4f4f-8d03-e028bb9210c9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.349368] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1044.349368] env[61824]: value = "task-1276074" [ 1044.349368] env[61824]: _type = "Task" [ 1044.349368] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.358148] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276074, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.412110] env[61824]: DEBUG nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1044.511785] env[61824]: DEBUG nova.network.neutron [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updated VIF entry in instance network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1044.512267] env[61824]: DEBUG nova.network.neutron [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.812336] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fb2d8d-fdc0-9eef-4259-253ab64e6484, 'name': SearchDatastore_Task, 'duration_secs': 0.007993} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.813238] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f92bd70a-212f-46ff-b354-089f30ce4adb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.818692] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1044.818692] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52fd967e-799c-b6bb-043d-79aff3d06c2a" [ 1044.818692] env[61824]: _type = "Task" [ 1044.818692] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.826697] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fd967e-799c-b6bb-043d-79aff3d06c2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.859419] env[61824]: DEBUG oslo_vmware.api [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276074, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126547} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.859714] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.859911] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1044.860105] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.881993] env[61824]: INFO nova.scheduler.client.report [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted allocations for instance 76303768-1bb1-4a82-8ca8-df301c19d02d [ 1044.962320] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1044.962641] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1044.962833] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1044.963063] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1044.963279] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1044.963452] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1044.963697] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1044.963906] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1044.964126] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1044.964340] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1044.964554] env[61824]: DEBUG nova.virt.hardware [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1044.965528] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7afe1f4f-e57c-4829-b83f-0c3d01916674 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.974303] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3988dac5-3251-4b6f-aafd-445e5b16b886 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.989468] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f6:a8:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '578e7e39-b803-41c6-890e-a1867602c4b9', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1044.997164] env[61824]: DEBUG oslo.service.loopingcall [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.997460] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1044.997688] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f2267686-cc4d-4afd-9f83-8f9b56a6b987 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.014799] env[61824]: DEBUG oslo_concurrency.lockutils [req-286a2639-2f05-440d-80a9-3a7de3c70b73 req-e2a673c6-e0c9-4589-ad0d-f1fb920a4e53 service nova] Releasing lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.016720] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.016720] env[61824]: value = "task-1276075" [ 1045.016720] env[61824]: _type = "Task" [ 1045.016720] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.024444] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276075, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.330056] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52fd967e-799c-b6bb-043d-79aff3d06c2a, 'name': SearchDatastore_Task, 'duration_secs': 0.010661} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.330457] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.330637] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/b8db876d-e430-4c5f-a439-0601dbf4598d.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1045.330912] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0f147372-1e17-4c57-8e7f-c339437d96b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.337644] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1045.337644] env[61824]: value = "task-1276076" [ 1045.337644] env[61824]: _type = "Task" [ 1045.337644] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.346944] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276076, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.386310] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.386802] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.387221] env[61824]: DEBUG nova.objects.instance [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'resources' on Instance uuid 76303768-1bb1-4a82-8ca8-df301c19d02d {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.421780] env[61824]: DEBUG nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1045.443816] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.444164] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.444414] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.444657] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.444844] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.445121] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.445392] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.445692] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.446018] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.446288] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.446488] env[61824]: DEBUG nova.virt.hardware [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.447443] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a33f62-894d-4a67-9130-59d341de2311 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.456390] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb76bcf4-32b8-4789-a709-7c6e94bcc006 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.470202] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.475926] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Creating folder: Project (22f414c6763743bc9babf056de498397). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1045.476290] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c8fe1d2e-cbec-4304-846d-f543928836e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.485609] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Created folder: Project (22f414c6763743bc9babf056de498397) in parent group-v274074. [ 1045.485823] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Creating folder: Instances. Parent ref: group-v274236. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1045.486092] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbb6c06e-9f13-4863-9d16-12cbca038817 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.494251] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Created folder: Instances in parent group-v274236. [ 1045.494498] env[61824]: DEBUG oslo.service.loopingcall [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.494696] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1045.494903] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6614dba-01a0-46da-bed9-96a9246f1e77 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.511237] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.511237] env[61824]: value = "task-1276079" [ 1045.511237] env[61824]: _type = "Task" [ 1045.511237] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.518670] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276079, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.526555] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276075, 'name': CreateVM_Task, 'duration_secs': 0.421809} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.526720] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.527430] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.527607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.527972] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1045.528253] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-933817fd-c6d0-453d-9aaa-9eda12c3ad89 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.532573] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1045.532573] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c9b267-75cd-7722-961b-e13708a30250" [ 1045.532573] env[61824]: _type = "Task" [ 1045.532573] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.542280] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c9b267-75cd-7722-961b-e13708a30250, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.704853] env[61824]: DEBUG nova.compute.manager [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Received event network-vif-unplugged-b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.704997] env[61824]: DEBUG oslo_concurrency.lockutils [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] Acquiring lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.705293] env[61824]: DEBUG oslo_concurrency.lockutils [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.705486] env[61824]: DEBUG oslo_concurrency.lockutils [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.705666] env[61824]: DEBUG nova.compute.manager [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] No waiting events found dispatching network-vif-unplugged-b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.705841] env[61824]: WARNING nova.compute.manager [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Received unexpected event network-vif-unplugged-b8687d0f-80a9-409d-895f-553d2a9bdc70 for instance with vm_state shelved_offloaded and task_state None. [ 1045.706020] env[61824]: DEBUG nova.compute.manager [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Received event network-changed-b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.706272] env[61824]: DEBUG nova.compute.manager [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Refreshing instance network info cache due to event network-changed-b8687d0f-80a9-409d-895f-553d2a9bdc70. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.706650] env[61824]: DEBUG oslo_concurrency.lockutils [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] Acquiring lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.706878] env[61824]: DEBUG oslo_concurrency.lockutils [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] Acquired lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.707102] env[61824]: DEBUG nova.network.neutron [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Refreshing network info cache for port b8687d0f-80a9-409d-895f-553d2a9bdc70 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.850854] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276076, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497479} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.851184] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/b8db876d-e430-4c5f-a439-0601dbf4598d.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1045.851422] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.851691] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-440d8163-6fc4-46e0-bfcd-44e11ae94e22 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.858276] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1045.858276] env[61824]: value = "task-1276080" [ 1045.858276] env[61824]: _type = "Task" [ 1045.858276] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.866295] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.890259] env[61824]: DEBUG nova.objects.instance [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'numa_topology' on Instance uuid 76303768-1bb1-4a82-8ca8-df301c19d02d {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.021071] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276079, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.042840] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c9b267-75cd-7722-961b-e13708a30250, 'name': SearchDatastore_Task, 'duration_secs': 0.009773} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.043180] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.043421] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.043660] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.043809] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.043989] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.044269] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ac6e3bf-b9b5-4dbb-b5ce-ff49b90f4c41 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.052632] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.052826] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1046.053585] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88bde06d-02fb-4cdd-b56b-cf9438ebf621 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.058921] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1046.058921] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5238ef14-051e-a573-13b1-a2cae7a89538" [ 1046.058921] env[61824]: _type = "Task" [ 1046.058921] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.070449] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5238ef14-051e-a573-13b1-a2cae7a89538, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.368572] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069109} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.368968] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.369667] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65975cca-15e5-430e-b329-f9f8d824b443 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.392588] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/b8db876d-e430-4c5f-a439-0601dbf4598d.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.393166] env[61824]: DEBUG nova.objects.base [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Object Instance<76303768-1bb1-4a82-8ca8-df301c19d02d> lazy-loaded attributes: resources,numa_topology {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1046.395758] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b32dcee-2d4f-4896-b11b-2ca4478e49f2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.418566] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1046.418566] env[61824]: value = "task-1276081" [ 1046.418566] env[61824]: _type = "Task" [ 1046.418566] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.427192] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276081, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.453475] env[61824]: DEBUG nova.network.neutron [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Updated VIF entry in instance network info cache for port b8687d0f-80a9-409d-895f-553d2a9bdc70. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.453922] env[61824]: DEBUG nova.network.neutron [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Updating instance_info_cache with network_info: [{"id": "b8687d0f-80a9-409d-895f-553d2a9bdc70", "address": "fa:16:3e:4f:91:34", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": null, "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb8687d0f-80", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.519527] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a514ff-012f-4812-be3d-0735e5f40556 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.525182] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276079, 'name': CreateVM_Task, 'duration_secs': 0.518499} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.525864] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1046.526364] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.526535] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.526914] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1046.527221] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6698a06-b827-4003-99d2-68776bc4299c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.533636] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5732b09-113a-4b38-8b97-c52df39b57cd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.538366] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1046.538366] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c52e1e-0d94-c858-2421-743d683f7d49" [ 1046.538366] env[61824]: _type = "Task" [ 1046.538366] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.568383] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "76303768-1bb1-4a82-8ca8-df301c19d02d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.572903] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d99d599-1e76-4d11-834f-61b40730e5d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.579071] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c52e1e-0d94-c858-2421-743d683f7d49, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.580015] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.580293] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.580529] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.585278] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5238ef14-051e-a573-13b1-a2cae7a89538, 'name': SearchDatastore_Task, 'duration_secs': 0.01394} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.587970] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d51119b-eb04-4707-a5d4-fe088ef24591 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.591246] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dbe441-d1eb-4a54-946c-47d3e534b6d9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.598865] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1046.598865] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]520fbbad-83a0-a899-4c08-aa19755fe968" [ 1046.598865] env[61824]: _type = "Task" [ 1046.598865] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.607552] env[61824]: DEBUG nova.compute.provider_tree [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.617270] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520fbbad-83a0-a899-4c08-aa19755fe968, 'name': SearchDatastore_Task, 'duration_secs': 0.009078} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.618175] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.618467] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1046.618754] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.618963] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.619211] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0dcb600-19dc-4964-a20c-6a0dbe18ed9d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.621772] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1f339a0-6348-4307-92e3-ea4547042f01 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.628340] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1046.628340] env[61824]: value = "task-1276082" [ 1046.628340] env[61824]: _type = "Task" [ 1046.628340] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.632952] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.633165] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1046.633883] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-360e0a56-acb8-4d8b-9d8d-6051ddbca4b8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.641076] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276082, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.642358] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1046.642358] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5224c05c-93aa-1278-a9f5-60378f7b1f2e" [ 1046.642358] env[61824]: _type = "Task" [ 1046.642358] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.649897] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5224c05c-93aa-1278-a9f5-60378f7b1f2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.930692] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276081, 'name': ReconfigVM_Task, 'duration_secs': 0.343009} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.931017] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/b8db876d-e430-4c5f-a439-0601dbf4598d.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.931672] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b30ae4be-6f30-43ee-adf9-63d6a3dd0633 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.939077] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1046.939077] env[61824]: value = "task-1276083" [ 1046.939077] env[61824]: _type = "Task" [ 1046.939077] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.947759] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276083, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.956482] env[61824]: DEBUG oslo_concurrency.lockutils [req-ff9def32-e9fe-4246-9f7b-3155217a8826 req-0fb9a6f2-be47-48a9-b61e-b5df73360370 service nova] Releasing lock "refresh_cache-76303768-1bb1-4a82-8ca8-df301c19d02d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.113885] env[61824]: DEBUG nova.scheduler.client.report [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.141779] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276082, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.153822] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5224c05c-93aa-1278-a9f5-60378f7b1f2e, 'name': SearchDatastore_Task, 'duration_secs': 0.011882} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.154676] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f0cb546-a5be-4c84-a923-cf5e10b6352d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.160596] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1047.160596] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]520e889c-e3b8-8585-f748-6cd4356303d5" [ 1047.160596] env[61824]: _type = "Task" [ 1047.160596] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.170274] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520e889c-e3b8-8585-f748-6cd4356303d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.449587] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276083, 'name': Rename_Task, 'duration_secs': 0.153133} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.449975] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.450217] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7694891-20ee-43a0-a85e-ac28fbec5021 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.457935] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1047.457935] env[61824]: value = "task-1276084" [ 1047.457935] env[61824]: _type = "Task" [ 1047.457935] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.466198] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276084, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.619505] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.233s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.639692] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276082, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.76511} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.640013] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1047.640248] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1047.640514] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3206500c-2329-4022-a281-b79bdccf6e4b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.649038] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1047.649038] env[61824]: value = "task-1276085" [ 1047.649038] env[61824]: _type = "Task" [ 1047.649038] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.658105] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276085, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.672560] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520e889c-e3b8-8585-f748-6cd4356303d5, 'name': SearchDatastore_Task, 'duration_secs': 0.06449} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.672875] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.673178] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1047.673458] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c018a84-3541-4cc6-8f91-8c047394fce5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.681817] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1047.681817] env[61824]: value = "task-1276086" [ 1047.681817] env[61824]: _type = "Task" [ 1047.681817] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.690720] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276086, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.970601] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276084, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.130072] env[61824]: DEBUG oslo_concurrency.lockutils [None req-327c125a-576f-41bf-ba80-953cfc9a136d tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.462s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.131180] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.563s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.131422] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.131633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.131842] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.133956] env[61824]: INFO nova.compute.manager [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Terminating instance [ 1048.136052] env[61824]: DEBUG nova.compute.manager [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1048.136305] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.136656] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e0a9614-3484-4815-a7d0-cfe810a2b0b4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.151163] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef97078-f4ac-46c5-b9c5-9b2342793779 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.173987] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276085, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.191461} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.174287] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.175145] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14bf6cd9-6040-4cfe-b16f-a4a9b11c23da {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.187726] env[61824]: WARNING nova.virt.vmwareapi.vmops [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 76303768-1bb1-4a82-8ca8-df301c19d02d could not be found. [ 1048.188012] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1048.188296] env[61824]: INFO nova.compute.manager [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1048.188585] env[61824]: DEBUG oslo.service.loopingcall [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.202523] env[61824]: DEBUG nova.compute.manager [-] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.202681] env[61824]: DEBUG nova.network.neutron [-] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1048.213127] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.213424] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d71ae8a-ca9a-4b82-9257-d5ed098276fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.233733] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276086, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500561} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.235059] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1048.235312] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.235653] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1048.235653] env[61824]: value = "task-1276087" [ 1048.235653] env[61824]: _type = "Task" [ 1048.235653] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.235848] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-832e52eb-a2a1-461b-8956-5cbe93ca10c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.248181] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276087, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.249506] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1048.249506] env[61824]: value = "task-1276088" [ 1048.249506] env[61824]: _type = "Task" [ 1048.249506] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.259261] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276088, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.468835] env[61824]: DEBUG oslo_vmware.api [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276084, 'name': PowerOnVM_Task, 'duration_secs': 0.705199} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.469226] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.469378] env[61824]: INFO nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Took 7.44 seconds to spawn the instance on the hypervisor. [ 1048.469575] env[61824]: DEBUG nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.470402] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8032d95-2760-4a60-9ca8-d400e7cde4a8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.748755] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276087, 'name': ReconfigVM_Task, 'duration_secs': 0.330535} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.749890] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Reconfigured VM instance instance-00000067 to attach disk [datastore2] a44b2d44-ecab-412a-b22e-c5a720b59631/a44b2d44-ecab-412a-b22e-c5a720b59631.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1048.749890] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f426bcce-af00-42a5-b9b8-765898865b91 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.759602] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276088, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071463} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.760703] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.761105] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1048.761105] env[61824]: value = "task-1276089" [ 1048.761105] env[61824]: _type = "Task" [ 1048.761105] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.762064] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b2e8473-7052-4cc2-bc12-751d3e39f467 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.773573] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276089, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.790552] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.791271] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b78c95e0-4be5-4a15-a880-5badf1a6292d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.811387] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1048.811387] env[61824]: value = "task-1276090" [ 1048.811387] env[61824]: _type = "Task" [ 1048.811387] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.819968] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276090, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.990986] env[61824]: INFO nova.compute.manager [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Took 12.70 seconds to build instance. [ 1049.050527] env[61824]: DEBUG nova.network.neutron [-] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.273647] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276089, 'name': Rename_Task, 'duration_secs': 0.153954} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.274090] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.274423] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f2f3221-745b-401f-99b3-438f1b962232 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.280487] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1049.280487] env[61824]: value = "task-1276091" [ 1049.280487] env[61824]: _type = "Task" [ 1049.280487] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.288116] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276091, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.322520] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276090, 'name': ReconfigVM_Task, 'duration_secs': 0.295852} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.322800] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.323839] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9c1a3d9-1d9e-4aca-928d-340fb321242f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.330508] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1049.330508] env[61824]: value = "task-1276092" [ 1049.330508] env[61824]: _type = "Task" [ 1049.330508] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.338772] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276092, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.493407] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8724fab7-faa4-4d76-9834-0397482d2007 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.207s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.553512] env[61824]: INFO nova.compute.manager [-] [instance: 76303768-1bb1-4a82-8ca8-df301c19d02d] Took 1.35 seconds to deallocate network for instance. [ 1049.791915] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276091, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.841580] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276092, 'name': Rename_Task, 'duration_secs': 0.164813} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.841908] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.843269] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd7ad206-f603-4e91-8cff-ce1c1aaa44c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.851068] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1049.851068] env[61824]: value = "task-1276093" [ 1049.851068] env[61824]: _type = "Task" [ 1049.851068] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.859355] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.292374] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276091, 'name': PowerOnVM_Task, 'duration_secs': 0.731542} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.292680] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.292917] env[61824]: DEBUG nova.compute.manager [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1050.293754] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c77cb73-9ba2-4148-a9ad-1dee7b0b2fab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.362676] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276093, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.434891] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b914c-fbb0-4f20-10f9-00c1fb456f9f/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1050.435890] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18d7fa52-874b-4d45-a6ba-5fe6c7743372 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.442524] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b914c-fbb0-4f20-10f9-00c1fb456f9f/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1050.442758] env[61824]: ERROR oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b914c-fbb0-4f20-10f9-00c1fb456f9f/disk-0.vmdk due to incomplete transfer. [ 1050.442989] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a58819be-71fc-478a-a16d-4591e1b04c9c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.450615] env[61824]: DEBUG oslo_vmware.rw_handles [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/526b914c-fbb0-4f20-10f9-00c1fb456f9f/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1050.451015] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Uploaded image c263e67c-1047-4baa-835b-62b408c6dbe8 to the Glance image server {{(pid=61824) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1050.454164] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Destroying the VM {{(pid=61824) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1050.454456] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b8810339-543b-4a9c-a14c-1303b591ff2f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.462021] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1050.462021] env[61824]: value = "task-1276094" [ 1050.462021] env[61824]: _type = "Task" [ 1050.462021] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.471978] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276094, 'name': Destroy_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.580883] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f69b9ba-f342-4692-9418-5ee1ca202fb7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "76303768-1bb1-4a82-8ca8-df301c19d02d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.450s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.643486] env[61824]: DEBUG nova.compute.manager [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Received event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1050.643701] env[61824]: DEBUG nova.compute.manager [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing instance network info cache due to event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1050.643924] env[61824]: DEBUG oslo_concurrency.lockutils [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] Acquiring lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.644086] env[61824]: DEBUG oslo_concurrency.lockutils [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] Acquired lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.644256] env[61824]: DEBUG nova.network.neutron [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1050.807024] env[61824]: INFO nova.compute.manager [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] bringing vm to original state: 'stopped' [ 1050.861246] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276093, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.973343] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276094, 'name': Destroy_Task} progress is 33%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.361818] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276093, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.410177] env[61824]: DEBUG nova.network.neutron [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updated VIF entry in instance network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1051.410556] env[61824]: DEBUG nova.network.neutron [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.472886] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276094, 'name': Destroy_Task, 'duration_secs': 0.733671} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.473190] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Destroyed the VM [ 1051.473428] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Deleting Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1051.473739] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ea530332-ddb7-4c87-99fd-23dabb1ec592 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.480790] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1051.480790] env[61824]: value = "task-1276095" [ 1051.480790] env[61824]: _type = "Task" [ 1051.480790] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.488530] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276095, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.644095] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.644406] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.812444] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "a44b2d44-ecab-412a-b22e-c5a720b59631" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.812748] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.812958] env[61824]: DEBUG nova.compute.manager [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1051.813899] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84b865d-54d0-49ba-9ded-293b0a5de690 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.822733] env[61824]: DEBUG nova.compute.manager [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1051.824777] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1051.825010] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efa1a544-7115-4b2b-9b68-f4fed4104b6f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.832280] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1051.832280] env[61824]: value = "task-1276096" [ 1051.832280] env[61824]: _type = "Task" [ 1051.832280] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.840466] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.861417] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276093, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.913123] env[61824]: DEBUG oslo_concurrency.lockutils [req-d9162ff6-fed0-495a-a479-9c42e8a23386 req-d0a9d00d-d58b-40b3-bb2b-ae25c6479648 service nova] Releasing lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.992110] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276095, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.147243] env[61824]: DEBUG nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1052.343324] env[61824]: DEBUG oslo_vmware.api [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276096, 'name': PowerOffVM_Task, 'duration_secs': 0.178882} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.343607] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1052.343788] env[61824]: DEBUG nova.compute.manager [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.344589] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe245ea6-1d42-428a-a607-16762e6f74bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.361800] env[61824]: DEBUG oslo_vmware.api [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276093, 'name': PowerOnVM_Task, 'duration_secs': 2.194281} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.362096] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.362319] env[61824]: INFO nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Took 6.94 seconds to spawn the instance on the hypervisor. [ 1052.362509] env[61824]: DEBUG nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.363329] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a42ee7-32a0-4209-bac9-9d449fc8fbf8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.491251] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276095, 'name': RemoveSnapshot_Task, 'duration_secs': 0.916531} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.491525] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Deleted Snapshot of the VM instance {{(pid=61824) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1052.491817] env[61824]: DEBUG nova.compute.manager [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.492602] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579908f8-580f-4e21-9a86-22df5e694538 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.669475] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.669736] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.671242] env[61824]: INFO nova.compute.claims [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.859863] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.047s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.880881] env[61824]: INFO nova.compute.manager [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Took 11.67 seconds to build instance. [ 1053.003822] env[61824]: INFO nova.compute.manager [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Shelve offloading [ 1053.005561] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1053.005810] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b26fbaac-353f-4574-817f-c2715bf53936 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.013458] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1053.013458] env[61824]: value = "task-1276097" [ 1053.013458] env[61824]: _type = "Task" [ 1053.013458] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.022330] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276097, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.367013] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.382873] env[61824]: DEBUG oslo_concurrency.lockutils [None req-146019ae-f4f1-484a-9c9e-827598f558ea tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "358cd7ec-99b4-446c-9e60-2fe50d375893" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.181s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.524714] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1053.525236] env[61824]: DEBUG nova.compute.manager [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.525693] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cf7bea-ddea-4c6d-8bf3-b86c9f5135c1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.531416] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.531596] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.531769] env[61824]: DEBUG nova.network.neutron [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1053.774291] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b630e911-7fbb-4dfb-9447-4647ae6624c8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.782311] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fb9836-ca2f-4533-8a5f-d5b56847a42c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.812792] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdab5d60-acdc-4beb-9404-3200fb8eb59d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.823088] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33a0365-7102-48e6-a5f5-04720c6bd9de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.838791] env[61824]: DEBUG nova.compute.provider_tree [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.843023] env[61824]: INFO nova.compute.manager [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Rebuilding instance [ 1053.877792] env[61824]: DEBUG nova.compute.manager [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1053.878686] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15794e1e-b471-4905-aa85-d23a74484e2d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.229854] env[61824]: DEBUG nova.network.neutron [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c770083-4d", "ovs_interfaceid": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.318624] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "a44b2d44-ecab-412a-b22e-c5a720b59631" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.318901] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.319145] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "a44b2d44-ecab-412a-b22e-c5a720b59631-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.319343] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.319514] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.321966] env[61824]: INFO nova.compute.manager [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Terminating instance [ 1054.323639] env[61824]: DEBUG nova.compute.manager [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1054.323832] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.324659] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7ccc40-9212-43c5-b3d6-ca63938b7203 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.332533] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.332768] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d60e5ff-7a8b-4697-a3d4-b6041f3b9d04 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.341708] env[61824]: DEBUG nova.scheduler.client.report [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.389273] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.389616] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba4e627d-d9e3-462d-8b42-a7ba61cc55fd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.399048] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1054.399048] env[61824]: value = "task-1276099" [ 1054.399048] env[61824]: _type = "Task" [ 1054.399048] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.409091] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.413586] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.413813] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.414128] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleting the datastore file [datastore2] a44b2d44-ecab-412a-b22e-c5a720b59631 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.414341] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5f33b39-5ed5-489d-873c-c4fa1b0ba318 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.422596] env[61824]: DEBUG oslo_vmware.api [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1054.422596] env[61824]: value = "task-1276100" [ 1054.422596] env[61824]: _type = "Task" [ 1054.422596] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.431174] env[61824]: DEBUG oslo_vmware.api [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276100, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.732639] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.846745] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.847299] env[61824]: DEBUG nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1054.850415] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.484s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.850604] env[61824]: DEBUG nova.objects.instance [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1054.909374] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276099, 'name': PowerOffVM_Task, 'duration_secs': 0.182945} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.909646] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1054.909903] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.910702] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21af772f-fb7f-4e61-833a-b0bac388151d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.918026] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1054.918552] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ec13646c-19d4-4c51-b83a-3912a59a7572 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.934560] env[61824]: DEBUG oslo_vmware.api [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276100, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133678} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.934900] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1054.935166] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1054.935485] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1054.935692] env[61824]: INFO nova.compute.manager [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1054.935946] env[61824]: DEBUG oslo.service.loopingcall [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.936198] env[61824]: DEBUG nova.compute.manager [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1054.936315] env[61824]: DEBUG nova.network.neutron [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1054.948295] env[61824]: DEBUG nova.compute.manager [req-05cedf70-5b6f-429d-9e4a-2591e8d81bfb req-a7d28c54-ca67-41e8-8305-1241221b3141 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-vif-unplugged-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.948530] env[61824]: DEBUG oslo_concurrency.lockutils [req-05cedf70-5b6f-429d-9e4a-2591e8d81bfb req-a7d28c54-ca67-41e8-8305-1241221b3141 service nova] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.948750] env[61824]: DEBUG oslo_concurrency.lockutils [req-05cedf70-5b6f-429d-9e4a-2591e8d81bfb req-a7d28c54-ca67-41e8-8305-1241221b3141 service nova] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.948916] env[61824]: DEBUG oslo_concurrency.lockutils [req-05cedf70-5b6f-429d-9e4a-2591e8d81bfb req-a7d28c54-ca67-41e8-8305-1241221b3141 service nova] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.949104] env[61824]: DEBUG nova.compute.manager [req-05cedf70-5b6f-429d-9e4a-2591e8d81bfb req-a7d28c54-ca67-41e8-8305-1241221b3141 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] No waiting events found dispatching network-vif-unplugged-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1054.949283] env[61824]: WARNING nova.compute.manager [req-05cedf70-5b6f-429d-9e4a-2591e8d81bfb req-a7d28c54-ca67-41e8-8305-1241221b3141 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received unexpected event network-vif-unplugged-4c770083-4d53-46bb-b9a4-cdaa981b457c for instance with vm_state shelved and task_state shelving_offloading. [ 1054.955342] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1054.955561] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1054.955750] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Deleting the datastore file [datastore2] 358cd7ec-99b4-446c-9e60-2fe50d375893 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1054.956015] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a33035f-676a-42ae-9ece-cdcf5ab212a4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.963779] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1054.963779] env[61824]: value = "task-1276102" [ 1054.963779] env[61824]: _type = "Task" [ 1054.963779] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.973980] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276102, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.186933] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.187949] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42b8e64-31e8-493b-84dc-4aff46806cd0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.196303] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1055.196560] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2ff5c89-dac2-4e47-b04b-4c976181ae02 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.285377] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1055.285614] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1055.285803] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleting the datastore file [datastore1] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1055.286170] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c617e55-c793-4b5e-b3d5-f461931a6733 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.293774] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1055.293774] env[61824]: value = "task-1276104" [ 1055.293774] env[61824]: _type = "Task" [ 1055.293774] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.303311] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276104, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.355041] env[61824]: DEBUG nova.compute.utils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.360289] env[61824]: DEBUG nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1055.360482] env[61824]: DEBUG nova.network.neutron [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1055.401479] env[61824]: DEBUG nova.policy [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8763f64818e24fad80ce232ac7d4463f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '629276280a234e7a8b810e6e74779fff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1055.475474] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276102, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098808} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.475772] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.475979] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1055.476176] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.647122] env[61824]: DEBUG nova.network.neutron [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Successfully created port: 3851ef8b-da22-42dc-8365-a12e16af0f91 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.773635] env[61824]: DEBUG nova.network.neutron [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.808449] env[61824]: DEBUG oslo_vmware.api [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276104, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12642} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.808729] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1055.808924] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1055.809198] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1055.841176] env[61824]: INFO nova.scheduler.client.report [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted allocations for instance 5bf9f3b5-7084-4c5a-952b-d90c337ad10a [ 1055.862176] env[61824]: DEBUG oslo_concurrency.lockutils [None req-07a4bbf5-c86e-4062-a422-b7517ba954d5 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.864316] env[61824]: DEBUG nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1056.277069] env[61824]: INFO nova.compute.manager [-] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Took 1.34 seconds to deallocate network for instance. [ 1056.345266] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.345550] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.345779] env[61824]: DEBUG nova.objects.instance [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'resources' on Instance uuid 5bf9f3b5-7084-4c5a-952b-d90c337ad10a {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.512790] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.513056] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.513226] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.513428] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.513579] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.513728] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.513941] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.514123] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.514296] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.514464] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.514641] env[61824]: DEBUG nova.virt.hardware [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.515547] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d929b900-111d-4c91-8963-3771d3e4fea6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.524640] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6d22f7-7e73-45c9-aa5b-45f411a8af82 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.541224] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Instance VIF info [] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.548781] env[61824]: DEBUG oslo.service.loopingcall [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.548781] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1056.548781] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b67c3742-2295-457c-83b7-4ebd318eaf19 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.568915] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.568915] env[61824]: value = "task-1276105" [ 1056.568915] env[61824]: _type = "Task" [ 1056.568915] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.577778] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276105, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.784100] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.848068] env[61824]: DEBUG nova.objects.instance [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'numa_topology' on Instance uuid 5bf9f3b5-7084-4c5a-952b-d90c337ad10a {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.871926] env[61824]: DEBUG nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1056.892521] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.892815] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.892933] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.893177] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.893258] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.893404] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.893621] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.893782] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.893952] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.894137] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.894318] env[61824]: DEBUG nova.virt.hardware [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.895436] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0e4237-642d-425b-a3e2-b122b6564df7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.904129] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0be80a-686c-4b57-8af2-f087140e43cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.978684] env[61824]: DEBUG nova.compute.manager [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1056.978894] env[61824]: DEBUG nova.compute.manager [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing instance network info cache due to event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1056.979132] env[61824]: DEBUG oslo_concurrency.lockutils [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] Acquiring lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.979285] env[61824]: DEBUG oslo_concurrency.lockutils [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] Acquired lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.979455] env[61824]: DEBUG nova.network.neutron [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1057.079231] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276105, 'name': CreateVM_Task, 'duration_secs': 0.295486} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.079401] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1057.079962] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.080063] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.080363] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.080617] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f5007b9-c782-41c6-88a9-dd95df5d3604 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.085133] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1057.085133] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5249dfed-bc8a-eeb0-8a36-6e3b1e2598be" [ 1057.085133] env[61824]: _type = "Task" [ 1057.085133] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.092779] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5249dfed-bc8a-eeb0-8a36-6e3b1e2598be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.350349] env[61824]: DEBUG nova.objects.base [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Object Instance<5bf9f3b5-7084-4c5a-952b-d90c337ad10a> lazy-loaded attributes: resources,numa_topology {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1057.400923] env[61824]: DEBUG nova.network.neutron [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Successfully updated port: 3851ef8b-da22-42dc-8365-a12e16af0f91 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.460999] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe71caee-b352-4aef-8524-c70582cc7fc0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.468897] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57d9e55-2876-405e-883e-e4a9655a3ba0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.504075] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9dc260-3626-4d82-a8ed-55e13f2282a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.512165] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6acf6303-aac9-4ff0-bb77-d375628fc191 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.525634] env[61824]: DEBUG nova.compute.provider_tree [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.595755] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5249dfed-bc8a-eeb0-8a36-6e3b1e2598be, 'name': SearchDatastore_Task, 'duration_secs': 0.011017} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.596082] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.596346] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.596577] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.596730] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.596913] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.597205] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e0df0617-570a-4a58-bfb0-07f2d1172800 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.605785] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.605973] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1057.606698] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90928ec1-f32c-4e98-a393-6fa8431e54ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.612148] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1057.612148] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5244cc6b-09ec-f036-3d2a-8f4b3652f118" [ 1057.612148] env[61824]: _type = "Task" [ 1057.612148] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.623835] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5244cc6b-09ec-f036-3d2a-8f4b3652f118, 'name': SearchDatastore_Task, 'duration_secs': 0.009011} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.624622] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-007e7b44-871c-468b-9fed-2d8cd50c1dea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.630033] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1057.630033] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521eb7f1-b3f3-6e68-61d6-24d124cb774a" [ 1057.630033] env[61824]: _type = "Task" [ 1057.630033] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.637883] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521eb7f1-b3f3-6e68-61d6-24d124cb774a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.763208] env[61824]: DEBUG nova.network.neutron [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updated VIF entry in instance network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1057.763589] env[61824]: DEBUG nova.network.neutron [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4c770083-4d", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.903341] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-a83ebb06-0343-4a84-a71f-a84e47f9ec5e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.903571] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-a83ebb06-0343-4a84-a71f-a84e47f9ec5e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.903644] env[61824]: DEBUG nova.network.neutron [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1058.028168] env[61824]: DEBUG nova.scheduler.client.report [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.141358] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521eb7f1-b3f3-6e68-61d6-24d124cb774a, 'name': SearchDatastore_Task, 'duration_secs': 0.009361} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.141626] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.142050] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1058.142200] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2abc65a5-dd25-4527-906f-b890633bc7a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.151212] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1058.151212] env[61824]: value = "task-1276106" [ 1058.151212] env[61824]: _type = "Task" [ 1058.151212] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.161513] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.266288] env[61824]: DEBUG oslo_concurrency.lockutils [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] Releasing lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.266739] env[61824]: DEBUG nova.compute.manager [req-06913f57-ea35-435e-8e36-d7d4797560b8 req-83285dcb-ead5-4eb7-b97e-ed9c9c1f7f58 service nova] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Received event network-vif-deleted-578e7e39-b803-41c6-890e-a1867602c4b9 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.440913] env[61824]: DEBUG nova.network.neutron [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1058.532653] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.187s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.535272] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.751s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.535508] env[61824]: DEBUG nova.objects.instance [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'resources' on Instance uuid a44b2d44-ecab-412a-b22e-c5a720b59631 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.588658] env[61824]: DEBUG nova.network.neutron [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Updating instance_info_cache with network_info: [{"id": "3851ef8b-da22-42dc-8365-a12e16af0f91", "address": "fa:16:3e:1f:fc:ea", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3851ef8b-da", "ovs_interfaceid": "3851ef8b-da22-42dc-8365-a12e16af0f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.661737] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.436526} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.663178] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1058.663411] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1058.666482] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c7b8b98c-aa2e-4ea2-9dc3-1e7aaa1324e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.674932] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1058.674932] env[61824]: value = "task-1276107" [ 1058.674932] env[61824]: _type = "Task" [ 1058.674932] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.684468] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.007369] env[61824]: DEBUG nova.compute.manager [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Received event network-vif-plugged-3851ef8b-da22-42dc-8365-a12e16af0f91 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.007602] env[61824]: DEBUG oslo_concurrency.lockutils [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] Acquiring lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.007639] env[61824]: DEBUG oslo_concurrency.lockutils [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.007816] env[61824]: DEBUG oslo_concurrency.lockutils [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.008025] env[61824]: DEBUG nova.compute.manager [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] No waiting events found dispatching network-vif-plugged-3851ef8b-da22-42dc-8365-a12e16af0f91 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1059.008246] env[61824]: WARNING nova.compute.manager [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Received unexpected event network-vif-plugged-3851ef8b-da22-42dc-8365-a12e16af0f91 for instance with vm_state building and task_state spawning. [ 1059.008469] env[61824]: DEBUG nova.compute.manager [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Received event network-changed-3851ef8b-da22-42dc-8365-a12e16af0f91 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.008640] env[61824]: DEBUG nova.compute.manager [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Refreshing instance network info cache due to event network-changed-3851ef8b-da22-42dc-8365-a12e16af0f91. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.008817] env[61824]: DEBUG oslo_concurrency.lockutils [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] Acquiring lock "refresh_cache-a83ebb06-0343-4a84-a71f-a84e47f9ec5e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.043538] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4c54b0a2-9ca0-489b-84c7-cd30cfe14495 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.651s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.090989] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-a83ebb06-0343-4a84-a71f-a84e47f9ec5e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.091304] env[61824]: DEBUG nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Instance network_info: |[{"id": "3851ef8b-da22-42dc-8365-a12e16af0f91", "address": "fa:16:3e:1f:fc:ea", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3851ef8b-da", "ovs_interfaceid": "3851ef8b-da22-42dc-8365-a12e16af0f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1059.091658] env[61824]: DEBUG oslo_concurrency.lockutils [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] Acquired lock "refresh_cache-a83ebb06-0343-4a84-a71f-a84e47f9ec5e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.091866] env[61824]: DEBUG nova.network.neutron [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Refreshing network info cache for port 3851ef8b-da22-42dc-8365-a12e16af0f91 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1059.093073] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:fc:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3851ef8b-da22-42dc-8365-a12e16af0f91', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1059.100725] env[61824]: DEBUG oslo.service.loopingcall [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.103847] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1059.104257] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-470bec0c-1033-43e0-9079-949d0105fbdb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.129049] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1059.129049] env[61824]: value = "task-1276108" [ 1059.129049] env[61824]: _type = "Task" [ 1059.129049] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.142302] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276108, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.165224] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1559d4d5-5708-47db-b5de-8279eeccfde8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.169752] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.170124] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.170329] env[61824]: INFO nova.compute.manager [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Unshelving [ 1059.175422] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f106e4-9b80-4323-b957-9fca6732284c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.190082] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069743} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.216826] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1059.218320] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2899c5-24f4-4523-8129-3ca893bc3b42 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.221357] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce2ae06-c33a-4d9e-b2f7-eb2b654bda52 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.236410] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa08ddd-bf5f-4b49-bb7f-7ef37823001b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.248553] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.249133] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e744711-2c4c-4f76-a1c2-a8b26617f546 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.275591] env[61824]: DEBUG nova.compute.provider_tree [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.278071] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1059.278071] env[61824]: value = "task-1276109" [ 1059.278071] env[61824]: _type = "Task" [ 1059.278071] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.287973] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276109, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.641159] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276108, 'name': CreateVM_Task, 'duration_secs': 0.342095} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.641432] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1059.641998] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.642193] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.642520] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1059.642771] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa82766f-6665-4c03-ac1b-f6115c6692af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.647296] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1059.647296] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52732c0d-3b84-d282-1f6e-78855e203f74" [ 1059.647296] env[61824]: _type = "Task" [ 1059.647296] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.655083] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52732c0d-3b84-d282-1f6e-78855e203f74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.780106] env[61824]: DEBUG nova.scheduler.client.report [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1059.792758] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276109, 'name': ReconfigVM_Task, 'duration_secs': 0.282211} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.793050] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 358cd7ec-99b4-446c-9e60-2fe50d375893/358cd7ec-99b4-446c-9e60-2fe50d375893.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.793652] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0a07db0-42e6-463f-90ae-e3f93a763804 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.801269] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1059.801269] env[61824]: value = "task-1276110" [ 1059.801269] env[61824]: _type = "Task" [ 1059.801269] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.813613] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276110, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.859098] env[61824]: DEBUG nova.network.neutron [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Updated VIF entry in instance network info cache for port 3851ef8b-da22-42dc-8365-a12e16af0f91. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1059.859586] env[61824]: DEBUG nova.network.neutron [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Updating instance_info_cache with network_info: [{"id": "3851ef8b-da22-42dc-8365-a12e16af0f91", "address": "fa:16:3e:1f:fc:ea", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3851ef8b-da", "ovs_interfaceid": "3851ef8b-da22-42dc-8365-a12e16af0f91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.158088] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52732c0d-3b84-d282-1f6e-78855e203f74, 'name': SearchDatastore_Task, 'duration_secs': 0.012635} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.158427] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.158620] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1060.158852] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.159009] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.159196] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1060.159455] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5153ee6-dbc2-49f8-bb0d-97642e753850 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.168665] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1060.168836] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1060.169527] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-576e3adc-d98a-4869-9968-ed744768f4b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.174394] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1060.174394] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52c896ac-638d-5af5-aa4c-06b99527eac7" [ 1060.174394] env[61824]: _type = "Task" [ 1060.174394] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.184062] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c896ac-638d-5af5-aa4c-06b99527eac7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.194580] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.287991] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.752s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.290425] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.096s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.290668] env[61824]: DEBUG nova.objects.instance [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'pci_requests' on Instance uuid 5bf9f3b5-7084-4c5a-952b-d90c337ad10a {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.306973] env[61824]: INFO nova.scheduler.client.report [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted allocations for instance a44b2d44-ecab-412a-b22e-c5a720b59631 [ 1060.311457] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276110, 'name': Rename_Task, 'duration_secs': 0.133739} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.313838] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.314262] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8eac87dc-dce3-4c2e-862d-fb9dfbd727ae {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.322506] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1060.322506] env[61824]: value = "task-1276111" [ 1060.322506] env[61824]: _type = "Task" [ 1060.322506] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.330545] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276111, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.362662] env[61824]: DEBUG oslo_concurrency.lockutils [req-4c0e87af-be3a-495f-a68c-d83bef0f66de req-075bd9b3-21a4-4815-8ed3-d406bc9d7894 service nova] Releasing lock "refresh_cache-a83ebb06-0343-4a84-a71f-a84e47f9ec5e" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.685883] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52c896ac-638d-5af5-aa4c-06b99527eac7, 'name': SearchDatastore_Task, 'duration_secs': 0.022118} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.686829] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38fd95fa-8aab-4d31-a42e-3b3310cc4554 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.692894] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1060.692894] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]520b2fc2-1b02-7097-2bea-6c5a8fba8274" [ 1060.692894] env[61824]: _type = "Task" [ 1060.692894] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.701475] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520b2fc2-1b02-7097-2bea-6c5a8fba8274, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.794766] env[61824]: DEBUG nova.objects.instance [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'numa_topology' on Instance uuid 5bf9f3b5-7084-4c5a-952b-d90c337ad10a {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.817784] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e03587b6-fd31-456c-afb3-6b7c31a7827f tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "a44b2d44-ecab-412a-b22e-c5a720b59631" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.498s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.832646] env[61824]: DEBUG oslo_vmware.api [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276111, 'name': PowerOnVM_Task, 'duration_secs': 0.459729} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.833298] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1060.833379] env[61824]: DEBUG nova.compute.manager [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1060.834227] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75cc835-e292-4e36-8539-8a79615b171b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.203932] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520b2fc2-1b02-7097-2bea-6c5a8fba8274, 'name': SearchDatastore_Task, 'duration_secs': 0.009937} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.204256] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.204517] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] a83ebb06-0343-4a84-a71f-a84e47f9ec5e/a83ebb06-0343-4a84-a71f-a84e47f9ec5e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1061.204786] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb4f2e6d-a730-45fc-ae7c-81a18b4970f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.213884] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1061.213884] env[61824]: value = "task-1276112" [ 1061.213884] env[61824]: _type = "Task" [ 1061.213884] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.222738] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276112, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.297691] env[61824]: INFO nova.compute.claims [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1061.353773] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.725013] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276112, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465006} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.725013] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] a83ebb06-0343-4a84-a71f-a84e47f9ec5e/a83ebb06-0343-4a84-a71f-a84e47f9ec5e.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.725013] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.725013] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6a2b774-7584-4bcb-a663-522895d20c2f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.732217] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1061.732217] env[61824]: value = "task-1276113" [ 1061.732217] env[61824]: _type = "Task" [ 1061.732217] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.740026] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276113, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.242628] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276113, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064096} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.243053] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1062.243697] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d632dc68-0952-444d-b814-995881437069 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.265795] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] a83ebb06-0343-4a84-a71f-a84e47f9ec5e/a83ebb06-0343-4a84-a71f-a84e47f9ec5e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1062.266071] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06bd63d7-2592-4acd-a31d-09482ceee1a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.285782] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1062.285782] env[61824]: value = "task-1276114" [ 1062.285782] env[61824]: _type = "Task" [ 1062.285782] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.295118] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.402080] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa6f54e-83cb-4a14-8039-17d26d5fb8b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.409999] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348859c5-a9e2-40c8-9f85-86f019f8a54b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.441501] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ddc3c2c-eb0a-4490-8d9a-f229c11d0519 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.449376] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a9ac3f-e78c-4e40-a85a-d0d270865a69 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.462911] env[61824]: DEBUG nova.compute.provider_tree [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.511481] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "358cd7ec-99b4-446c-9e60-2fe50d375893" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.511744] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "358cd7ec-99b4-446c-9e60-2fe50d375893" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.511954] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "358cd7ec-99b4-446c-9e60-2fe50d375893-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.512164] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "358cd7ec-99b4-446c-9e60-2fe50d375893-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.512413] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "358cd7ec-99b4-446c-9e60-2fe50d375893-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.514491] env[61824]: INFO nova.compute.manager [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Terminating instance [ 1062.516107] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "refresh_cache-358cd7ec-99b4-446c-9e60-2fe50d375893" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.516274] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquired lock "refresh_cache-358cd7ec-99b4-446c-9e60-2fe50d375893" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.516440] env[61824]: DEBUG nova.network.neutron [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1062.560048] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.560048] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.795955] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276114, 'name': ReconfigVM_Task, 'duration_secs': 0.26326} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.796275] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Reconfigured VM instance instance-0000006a to attach disk [datastore1] a83ebb06-0343-4a84-a71f-a84e47f9ec5e/a83ebb06-0343-4a84-a71f-a84e47f9ec5e.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.796915] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc352bc7-3e99-47df-8fb2-75e67f607fe7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.803964] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1062.803964] env[61824]: value = "task-1276115" [ 1062.803964] env[61824]: _type = "Task" [ 1062.803964] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.811728] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276115, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.966979] env[61824]: DEBUG nova.scheduler.client.report [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1063.040395] env[61824]: DEBUG nova.network.neutron [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1063.061476] env[61824]: DEBUG nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1063.100517] env[61824]: DEBUG nova.network.neutron [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.314995] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276115, 'name': Rename_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.472042] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.181s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.474412] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.121s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.474509] env[61824]: DEBUG nova.objects.instance [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1063.505081] env[61824]: INFO nova.network.neutron [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating port 4c770083-4d53-46bb-b9a4-cdaa981b457c with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1063.584032] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.603346] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Releasing lock "refresh_cache-358cd7ec-99b4-446c-9e60-2fe50d375893" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.604178] env[61824]: DEBUG nova.compute.manager [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1063.604225] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1063.605464] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b071d015-efb0-4a80-9823-06ec8183b4c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.613720] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1063.614143] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a29937a8-ce2c-4402-9d35-7ed678c5a6ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.622715] env[61824]: DEBUG oslo_vmware.api [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1063.622715] env[61824]: value = "task-1276116" [ 1063.622715] env[61824]: _type = "Task" [ 1063.622715] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.632677] env[61824]: DEBUG oslo_vmware.api [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276116, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.815647] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276115, 'name': Rename_Task, 'duration_secs': 0.850368} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.815953] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.816219] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6d1dccd-107f-4f2d-8516-f1a4a5b8d808 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.822570] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1063.822570] env[61824]: value = "task-1276117" [ 1063.822570] env[61824]: _type = "Task" [ 1063.822570] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.829701] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276117, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.133738] env[61824]: DEBUG oslo_vmware.api [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276116, 'name': PowerOffVM_Task, 'duration_secs': 0.193126} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.134015] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1064.134204] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1064.134459] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bae69a86-eed1-44ba-a274-5557d98ebb2c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.162141] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1064.162380] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1064.162568] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Deleting the datastore file [datastore1] 358cd7ec-99b4-446c-9e60-2fe50d375893 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1064.162844] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edfec5f5-1c23-4245-a34e-838294c4d825 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.169874] env[61824]: DEBUG oslo_vmware.api [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for the task: (returnval){ [ 1064.169874] env[61824]: value = "task-1276119" [ 1064.169874] env[61824]: _type = "Task" [ 1064.169874] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.178142] env[61824]: DEBUG oslo_vmware.api [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276119, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.333172] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276117, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.483764] env[61824]: DEBUG oslo_concurrency.lockutils [None req-4a7bf8e2-5749-4d90-b782-13e5181b1d0e tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.485009] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.901s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.486471] env[61824]: INFO nova.compute.claims [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1064.681316] env[61824]: DEBUG oslo_vmware.api [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Task: {'id': task-1276119, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.107736} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.681545] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1064.681733] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1064.681912] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1064.682107] env[61824]: INFO nova.compute.manager [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1064.682361] env[61824]: DEBUG oslo.service.loopingcall [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.682584] env[61824]: DEBUG nova.compute.manager [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1064.682659] env[61824]: DEBUG nova.network.neutron [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1064.699337] env[61824]: DEBUG nova.network.neutron [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1064.834521] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276117, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.025216] env[61824]: DEBUG nova.compute.manager [req-da9d6ee3-8310-4b3b-95c7-2c5602993c4c req-23571841-4998-40c0-b8fd-867402d0746e service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-vif-plugged-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.025463] env[61824]: DEBUG oslo_concurrency.lockutils [req-da9d6ee3-8310-4b3b-95c7-2c5602993c4c req-23571841-4998-40c0-b8fd-867402d0746e service nova] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.025671] env[61824]: DEBUG oslo_concurrency.lockutils [req-da9d6ee3-8310-4b3b-95c7-2c5602993c4c req-23571841-4998-40c0-b8fd-867402d0746e service nova] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.025850] env[61824]: DEBUG oslo_concurrency.lockutils [req-da9d6ee3-8310-4b3b-95c7-2c5602993c4c req-23571841-4998-40c0-b8fd-867402d0746e service nova] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.026027] env[61824]: DEBUG nova.compute.manager [req-da9d6ee3-8310-4b3b-95c7-2c5602993c4c req-23571841-4998-40c0-b8fd-867402d0746e service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] No waiting events found dispatching network-vif-plugged-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1065.026214] env[61824]: WARNING nova.compute.manager [req-da9d6ee3-8310-4b3b-95c7-2c5602993c4c req-23571841-4998-40c0-b8fd-867402d0746e service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received unexpected event network-vif-plugged-4c770083-4d53-46bb-b9a4-cdaa981b457c for instance with vm_state shelved_offloaded and task_state spawning. [ 1065.124530] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.124802] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.124943] env[61824]: DEBUG nova.network.neutron [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1065.202096] env[61824]: DEBUG nova.network.neutron [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.334720] env[61824]: DEBUG oslo_vmware.api [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276117, 'name': PowerOnVM_Task, 'duration_secs': 1.485552} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.335106] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.335227] env[61824]: INFO nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Took 8.46 seconds to spawn the instance on the hypervisor. [ 1065.335416] env[61824]: DEBUG nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.336197] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b935e929-8cd7-44ef-ab8b-b9c9134d1535 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.603318] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64feba89-731b-420d-851d-70d0d6b752eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.611882] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28acaa43-048d-441d-90d7-1a39dcb0dd11 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.644665] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0178b6-41ae-4999-85ef-eee5216d80a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.652217] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f72406-84d1-47ec-ae6a-dbcf6e5b9769 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.665442] env[61824]: DEBUG nova.compute.provider_tree [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.703788] env[61824]: INFO nova.compute.manager [-] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Took 1.02 seconds to deallocate network for instance. [ 1065.852063] env[61824]: DEBUG nova.network.neutron [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c770083-4d", "ovs_interfaceid": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.855172] env[61824]: INFO nova.compute.manager [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Took 13.20 seconds to build instance. [ 1066.031033] env[61824]: DEBUG oslo_concurrency.lockutils [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.168748] env[61824]: DEBUG nova.scheduler.client.report [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.211874] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.355837] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.358395] env[61824]: DEBUG oslo_concurrency.lockutils [None req-a4c4bd18-bb35-4fcb-8316-ab553289857b tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.714s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.358675] env[61824]: DEBUG oslo_concurrency.lockutils [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.328s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.358916] env[61824]: DEBUG nova.compute.manager [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1066.359715] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346d3faf-1b0b-4932-821a-12ba5af87452 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.366794] env[61824]: DEBUG nova.compute.manager [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61824) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1066.367346] env[61824]: DEBUG nova.objects.instance [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'flavor' on Instance uuid a83ebb06-0343-4a84-a71f-a84e47f9ec5e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.380570] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='75b48b20faf6dc2d184d6c9b370e56e9',container_format='bare',created_at=2024-10-12T11:01:31Z,direct_url=,disk_format='vmdk',id=c263e67c-1047-4baa-835b-62b408c6dbe8,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1773515870-shelved',owner='8bd0e94b74d74f9898049c9cad364b5b',properties=ImageMetaProps,protected=,size=31665152,status='active',tags=,updated_at=2024-10-12T11:01:45Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1066.380796] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1066.380958] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1066.381189] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1066.381340] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1066.381510] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1066.381717] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1066.381880] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1066.382081] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1066.382279] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1066.382462] env[61824]: DEBUG nova.virt.hardware [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.383260] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e862311f-bbbc-4f50-996b-d124f702b71c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.392087] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cea8157-2364-4792-bd39-2c537c719fb3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.405837] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:b8:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '08e9585e-6186-4788-9fd9-24174ce45a6f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c770083-4d53-46bb-b9a4-cdaa981b457c', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.413132] env[61824]: DEBUG oslo.service.loopingcall [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.413377] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1066.413578] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f878be2b-9a05-445e-8fef-785c34aed3f9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.433316] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.433316] env[61824]: value = "task-1276120" [ 1066.433316] env[61824]: _type = "Task" [ 1066.433316] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.674836] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.190s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.675489] env[61824]: DEBUG nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1066.678550] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.467s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.679252] env[61824]: DEBUG nova.objects.instance [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lazy-loading 'resources' on Instance uuid 358cd7ec-99b4-446c-9e60-2fe50d375893 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.873053] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1066.873372] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1e699c5-fd4d-490c-b2d6-2a3a3e95420b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.881762] env[61824]: DEBUG oslo_vmware.api [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1066.881762] env[61824]: value = "task-1276121" [ 1066.881762] env[61824]: _type = "Task" [ 1066.881762] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.890028] env[61824]: DEBUG oslo_vmware.api [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.944122] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276120, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.048949] env[61824]: DEBUG nova.compute.manager [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.049183] env[61824]: DEBUG nova.compute.manager [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing instance network info cache due to event network-changed-4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1067.049407] env[61824]: DEBUG oslo_concurrency.lockutils [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] Acquiring lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.049558] env[61824]: DEBUG oslo_concurrency.lockutils [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] Acquired lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.049725] env[61824]: DEBUG nova.network.neutron [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Refreshing network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1067.181652] env[61824]: DEBUG nova.compute.utils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.186086] env[61824]: DEBUG nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1067.186086] env[61824]: DEBUG nova.network.neutron [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1067.233634] env[61824]: DEBUG nova.policy [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad4e8224518048e583d27b94721b3c22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd21256931aa54010b72beca4fb798f63', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1067.290052] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31886b58-8682-47d4-8d13-8cf458fc98c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.302024] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a982c01-c063-4378-8465-4d147e050a06 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.330807] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f50e44-34ca-4fc5-a472-47cefabaa96e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.338242] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab10c8c-b34c-4cb4-8c22-124060dc9503 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.351456] env[61824]: DEBUG nova.compute.provider_tree [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.391697] env[61824]: DEBUG oslo_vmware.api [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.450814] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276120, 'name': CreateVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.534492] env[61824]: DEBUG nova.network.neutron [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Successfully created port: 7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1067.686649] env[61824]: DEBUG nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1067.854423] env[61824]: DEBUG nova.scheduler.client.report [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1067.892982] env[61824]: DEBUG oslo_vmware.api [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276121, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.947877] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276120, 'name': CreateVM_Task, 'duration_secs': 1.398456} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.948101] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.949260] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.949472] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.949965] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.950238] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95dc3116-97f4-4736-b48c-3cd26177fa96 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.953283] env[61824]: DEBUG nova.network.neutron [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updated VIF entry in instance network info cache for port 4c770083-4d53-46bb-b9a4-cdaa981b457c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1067.953615] env[61824]: DEBUG nova.network.neutron [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [{"id": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "address": "fa:16:3e:6e:b8:d1", "network": {"id": "b8806735-297a-4a4c-b447-2314c3ff1b1a", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1680272701-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bd0e94b74d74f9898049c9cad364b5b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "08e9585e-6186-4788-9fd9-24174ce45a6f", "external-id": "nsx-vlan-transportzone-254", "segmentation_id": 254, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c770083-4d", "ovs_interfaceid": "4c770083-4d53-46bb-b9a4-cdaa981b457c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.958028] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1067.958028] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5258b141-f0ab-5bc1-bf83-2667400412ce" [ 1067.958028] env[61824]: _type = "Task" [ 1067.958028] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.963448] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5258b141-f0ab-5bc1-bf83-2667400412ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.359772] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.681s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.384039] env[61824]: INFO nova.scheduler.client.report [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Deleted allocations for instance 358cd7ec-99b4-446c-9e60-2fe50d375893 [ 1068.396977] env[61824]: DEBUG oslo_vmware.api [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276121, 'name': PowerOffVM_Task, 'duration_secs': 1.032215} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.397268] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1068.397268] env[61824]: DEBUG nova.compute.manager [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1068.398097] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6319057a-16cd-4bab-9d0a-07c577a0712a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.456678] env[61824]: DEBUG oslo_concurrency.lockutils [req-d0f4d724-eb06-4dd4-9014-24bf6980a0d9 req-baa3769d-cac4-4e97-8264-d38ab1884642 service nova] Releasing lock "refresh_cache-5bf9f3b5-7084-4c5a-952b-d90c337ad10a" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.466361] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.466614] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Processing image c263e67c-1047-4baa-835b-62b408c6dbe8 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.466853] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8/c263e67c-1047-4baa-835b-62b408c6dbe8.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.467008] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8/c263e67c-1047-4baa-835b-62b408c6dbe8.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.467201] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.467445] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce375276-7478-4c86-88ef-790c5c685570 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.475536] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.475712] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1068.476422] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2803be61-5e7a-4c89-91fb-7675b990771a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.481344] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1068.481344] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52bdb634-64e9-18ca-ebfb-254f14872b5e" [ 1068.481344] env[61824]: _type = "Task" [ 1068.481344] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.488627] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52bdb634-64e9-18ca-ebfb-254f14872b5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.696193] env[61824]: DEBUG nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1068.720425] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.720669] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.720832] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.721050] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.721216] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.721368] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.721575] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.721736] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.721903] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.722075] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.722254] env[61824]: DEBUG nova.virt.hardware [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.723127] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a90dcce-914a-40f8-bdc1-6a88ccb333af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.730961] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0544e4df-1673-4899-87e1-f264548f31e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.892489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-cc6b3292-db21-4314-90ca-95780838b20b tempest-ServerShowV254Test-1598595318 tempest-ServerShowV254Test-1598595318-project-member] Lock "358cd7ec-99b4-446c-9e60-2fe50d375893" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.381s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.908670] env[61824]: DEBUG oslo_concurrency.lockutils [None req-168b5dcd-6545-48a9-8013-cdc63c19cdae tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.550s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.992350] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Preparing fetch location {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1068.992610] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Fetch image to [datastore2] OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1/OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1.vmdk {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1068.992798] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Downloading stream optimized image c263e67c-1047-4baa-835b-62b408c6dbe8 to [datastore2] OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1/OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1.vmdk on the data store datastore2 as vApp {{(pid=61824) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1068.992991] env[61824]: DEBUG nova.virt.vmwareapi.images [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Downloading image file data c263e67c-1047-4baa-835b-62b408c6dbe8 to the ESX as VM named 'OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1' {{(pid=61824) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1069.060832] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1069.060832] env[61824]: value = "resgroup-9" [ 1069.060832] env[61824]: _type = "ResourcePool" [ 1069.060832] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1069.061304] env[61824]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-23baabd0-2723-4735-b0ac-bed3fddf464d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.081608] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease: (returnval){ [ 1069.081608] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52197b76-7e38-d37e-ca17-c82e59c0c422" [ 1069.081608] env[61824]: _type = "HttpNfcLease" [ 1069.081608] env[61824]: } obtained for vApp import into resource pool (val){ [ 1069.081608] env[61824]: value = "resgroup-9" [ 1069.081608] env[61824]: _type = "ResourcePool" [ 1069.081608] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1069.081921] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the lease: (returnval){ [ 1069.081921] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52197b76-7e38-d37e-ca17-c82e59c0c422" [ 1069.081921] env[61824]: _type = "HttpNfcLease" [ 1069.081921] env[61824]: } to be ready. {{(pid=61824) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1069.089553] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1069.089553] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52197b76-7e38-d37e-ca17-c82e59c0c422" [ 1069.089553] env[61824]: _type = "HttpNfcLease" [ 1069.089553] env[61824]: } is initializing. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1069.219879] env[61824]: DEBUG nova.compute.manager [req-8629dcb4-02e8-45a6-8385-174ffa471851 req-06f43832-5687-4eb4-ab7d-66ad105a6197 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Received event network-vif-plugged-7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.220180] env[61824]: DEBUG oslo_concurrency.lockutils [req-8629dcb4-02e8-45a6-8385-174ffa471851 req-06f43832-5687-4eb4-ab7d-66ad105a6197 service nova] Acquiring lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.220410] env[61824]: DEBUG oslo_concurrency.lockutils [req-8629dcb4-02e8-45a6-8385-174ffa471851 req-06f43832-5687-4eb4-ab7d-66ad105a6197 service nova] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.220678] env[61824]: DEBUG oslo_concurrency.lockutils [req-8629dcb4-02e8-45a6-8385-174ffa471851 req-06f43832-5687-4eb4-ab7d-66ad105a6197 service nova] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.220760] env[61824]: DEBUG nova.compute.manager [req-8629dcb4-02e8-45a6-8385-174ffa471851 req-06f43832-5687-4eb4-ab7d-66ad105a6197 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] No waiting events found dispatching network-vif-plugged-7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1069.220910] env[61824]: WARNING nova.compute.manager [req-8629dcb4-02e8-45a6-8385-174ffa471851 req-06f43832-5687-4eb4-ab7d-66ad105a6197 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Received unexpected event network-vif-plugged-7afd60b2-5996-4b14-9a2a-2c271dc1394c for instance with vm_state building and task_state spawning. [ 1069.368606] env[61824]: DEBUG nova.network.neutron [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Successfully updated port: 7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1069.590232] env[61824]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1069.590232] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52197b76-7e38-d37e-ca17-c82e59c0c422" [ 1069.590232] env[61824]: _type = "HttpNfcLease" [ 1069.590232] env[61824]: } is ready. {{(pid=61824) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1069.590564] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1069.590564] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52197b76-7e38-d37e-ca17-c82e59c0c422" [ 1069.590564] env[61824]: _type = "HttpNfcLease" [ 1069.590564] env[61824]: }. {{(pid=61824) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1069.591473] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b9bf21-9470-49ea-a8fb-3252096b2ef1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.598679] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cac8a-03e4-03cd-b2d1-463274a71448/disk-0.vmdk from lease info. {{(pid=61824) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1069.598923] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating HTTP connection to write to file with size = 31665152 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cac8a-03e4-03cd-b2d1-463274a71448/disk-0.vmdk. {{(pid=61824) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1069.662498] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-720d1ebc-cb66-46c6-8644-dc662e12413c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.701467] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.701740] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.701948] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.702161] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.702363] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.704557] env[61824]: INFO nova.compute.manager [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Terminating instance [ 1069.706394] env[61824]: DEBUG nova.compute.manager [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1069.706585] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1069.707417] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24f8b91-bc42-4b9e-ba50-200d2b1531d3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.715129] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.715370] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2d188548-4658-4a0f-aa37-f1160e082812 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.794245] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.795060] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.795060] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleting the datastore file [datastore1] a83ebb06-0343-4a84-a71f-a84e47f9ec5e {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.795060] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c6868339-d0cf-4ea9-813d-6f3eea45807b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.801601] env[61824]: DEBUG oslo_vmware.api [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1069.801601] env[61824]: value = "task-1276124" [ 1069.801601] env[61824]: _type = "Task" [ 1069.801601] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.811680] env[61824]: DEBUG oslo_vmware.api [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.871466] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.871684] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.871800] env[61824]: DEBUG nova.network.neutron [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1070.314310] env[61824]: DEBUG oslo_vmware.api [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145975} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.314808] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1070.314858] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1070.315086] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1070.315309] env[61824]: INFO nova.compute.manager [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1070.315601] env[61824]: DEBUG oslo.service.loopingcall [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.315844] env[61824]: DEBUG nova.compute.manager [-] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1070.315967] env[61824]: DEBUG nova.network.neutron [-] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1070.431229] env[61824]: DEBUG nova.network.neutron [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1070.731363] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Completed reading data from the image iterator. {{(pid=61824) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1070.731750] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cac8a-03e4-03cd-b2d1-463274a71448/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1070.733365] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95728e22-46ba-43a8-a299-93481a6d77d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.740015] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cac8a-03e4-03cd-b2d1-463274a71448/disk-0.vmdk is in state: ready. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1070.740289] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cac8a-03e4-03cd-b2d1-463274a71448/disk-0.vmdk. {{(pid=61824) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1070.740578] env[61824]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-fd72e8ac-5c3c-4ad2-9e4e-1fbedf60eb9f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.806143] env[61824]: DEBUG nova.network.neutron [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updating instance_info_cache with network_info: [{"id": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "address": "fa:16:3e:35:ae:d2", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7afd60b2-59", "ovs_interfaceid": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.974949] env[61824]: DEBUG oslo_vmware.rw_handles [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521cac8a-03e4-03cd-b2d1-463274a71448/disk-0.vmdk. {{(pid=61824) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1070.975201] env[61824]: INFO nova.virt.vmwareapi.images [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Downloaded image file data c263e67c-1047-4baa-835b-62b408c6dbe8 [ 1070.976019] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58db72cb-d7a3-4cb0-b396-f6ed33270cee {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.991452] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f76fd553-3139-4b14-adcc-e9c7306bd872 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.024394] env[61824]: INFO nova.virt.vmwareapi.images [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] The imported VM was unregistered [ 1071.026970] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Caching image {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1071.027239] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Creating directory with path [datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1071.027514] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23c69069-0e5a-448e-b59a-878d6bbf0384 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.037458] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Created directory with path [datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8 {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1071.037644] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1/OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1.vmdk to [datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8/c263e67c-1047-4baa-835b-62b408c6dbe8.vmdk. {{(pid=61824) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1071.037891] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-323d2e8c-ed2f-4628-afdb-ce6211e1ddd0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.044713] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1071.044713] env[61824]: value = "task-1276126" [ 1071.044713] env[61824]: _type = "Task" [ 1071.044713] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.052370] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276126, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.258965] env[61824]: DEBUG nova.compute.manager [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Received event network-changed-7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1071.259127] env[61824]: DEBUG nova.compute.manager [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Refreshing instance network info cache due to event network-changed-7afd60b2-5996-4b14-9a2a-2c271dc1394c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1071.259334] env[61824]: DEBUG oslo_concurrency.lockutils [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] Acquiring lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.308756] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.309236] env[61824]: DEBUG nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Instance network_info: |[{"id": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "address": "fa:16:3e:35:ae:d2", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7afd60b2-59", "ovs_interfaceid": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1071.309653] env[61824]: DEBUG oslo_concurrency.lockutils [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] Acquired lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.309894] env[61824]: DEBUG nova.network.neutron [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Refreshing network info cache for port 7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1071.311452] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:ae:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7afd60b2-5996-4b14-9a2a-2c271dc1394c', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.319628] env[61824]: DEBUG oslo.service.loopingcall [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.320829] env[61824]: DEBUG nova.network.neutron [-] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.322066] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1071.324673] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d356260-9768-4651-aeaf-0a315a1a7e67 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.347371] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.347371] env[61824]: value = "task-1276127" [ 1071.347371] env[61824]: _type = "Task" [ 1071.347371] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.356567] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276127, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.555049] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276126, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.824388] env[61824]: INFO nova.compute.manager [-] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Took 1.51 seconds to deallocate network for instance. [ 1071.857220] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276127, 'name': CreateVM_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.055349] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276126, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.073262] env[61824]: DEBUG nova.network.neutron [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updated VIF entry in instance network info cache for port 7afd60b2-5996-4b14-9a2a-2c271dc1394c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1072.073648] env[61824]: DEBUG nova.network.neutron [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updating instance_info_cache with network_info: [{"id": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "address": "fa:16:3e:35:ae:d2", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7afd60b2-59", "ovs_interfaceid": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.331725] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.331725] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.331928] env[61824]: DEBUG nova.objects.instance [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'resources' on Instance uuid a83ebb06-0343-4a84-a71f-a84e47f9ec5e {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.358694] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276127, 'name': CreateVM_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.556229] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276126, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.576193] env[61824]: DEBUG oslo_concurrency.lockutils [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] Releasing lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.576530] env[61824]: DEBUG nova.compute.manager [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Received event network-vif-deleted-3851ef8b-da22-42dc-8365-a12e16af0f91 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1072.576711] env[61824]: INFO nova.compute.manager [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Neutron deleted interface 3851ef8b-da22-42dc-8365-a12e16af0f91; detaching it from the instance and deleting it from the info cache [ 1072.576917] env[61824]: DEBUG nova.network.neutron [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.858848] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276127, 'name': CreateVM_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.948016] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b5f98c-e423-416a-a5a0-e3bf00c0da55 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.955927] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c396d766-d709-455c-a163-4d5ecb485499 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.988426] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dfd090-3730-4f62-8513-3af034b83080 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.996361] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4227f8ef-3998-4307-9096-d7329115319e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.011683] env[61824]: DEBUG nova.compute.provider_tree [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.056592] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276126, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.079576] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-015b3b15-bc22-4a54-8561-0bfb41c7054e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.088578] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd5ffcb-c367-4c6b-a3eb-406fefe5df50 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.115207] env[61824]: DEBUG nova.compute.manager [req-de5bf1bb-445d-4c6d-97d8-0087f53757d6 req-18da4e2f-450c-4f56-9787-c57de472a833 service nova] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Detach interface failed, port_id=3851ef8b-da22-42dc-8365-a12e16af0f91, reason: Instance a83ebb06-0343-4a84-a71f-a84e47f9ec5e could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1073.358161] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276127, 'name': CreateVM_Task, 'duration_secs': 1.922937} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.358303] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1073.358977] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.359163] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.359489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.359743] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c4bce78-3f78-47c6-918c-3c55741f9fc0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.363941] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1073.363941] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52785624-7610-c89f-a43c-f41ba68bcf37" [ 1073.363941] env[61824]: _type = "Task" [ 1073.363941] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.370944] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52785624-7610-c89f-a43c-f41ba68bcf37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.515352] env[61824]: DEBUG nova.scheduler.client.report [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1073.556344] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276126, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.282033} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.556637] env[61824]: INFO nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1/OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1.vmdk to [datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8/c263e67c-1047-4baa-835b-62b408c6dbe8.vmdk. [ 1073.556849] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Cleaning up location [datastore2] OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1073.557029] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_1fbd277f-83c7-495f-8cb7-ccf27e963aa1 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1073.557282] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79b836f8-5a56-4cfe-8310-2c2241843a8d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.563323] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1073.563323] env[61824]: value = "task-1276128" [ 1073.563323] env[61824]: _type = "Task" [ 1073.563323] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.571531] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276128, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.873845] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52785624-7610-c89f-a43c-f41ba68bcf37, 'name': SearchDatastore_Task, 'duration_secs': 0.009808} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.874167] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.874340] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.874579] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.874727] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.874907] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.875181] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcb25609-f575-4de5-9456-4ea306838701 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.882950] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.883139] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1073.883807] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b093afac-7b1f-4119-a424-bacabc2e2b43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.888241] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1073.888241] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5231b7ae-2046-1720-46a0-e5e1177021e9" [ 1073.888241] env[61824]: _type = "Task" [ 1073.888241] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.895991] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231b7ae-2046-1720-46a0-e5e1177021e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.020287] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.037017] env[61824]: INFO nova.scheduler.client.report [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted allocations for instance a83ebb06-0343-4a84-a71f-a84e47f9ec5e [ 1074.072231] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276128, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.038304} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.072488] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.072657] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8/c263e67c-1047-4baa-835b-62b408c6dbe8.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.072927] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8/c263e67c-1047-4baa-835b-62b408c6dbe8.vmdk to [datastore2] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.073221] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1f88614-ee9a-4e19-978d-2282b67f9eb5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.079606] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1074.079606] env[61824]: value = "task-1276129" [ 1074.079606] env[61824]: _type = "Task" [ 1074.079606] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.087795] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.399050] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231b7ae-2046-1720-46a0-e5e1177021e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009093} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.399896] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7540b9df-e79b-4877-802d-22f8445fb030 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.405246] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1074.405246] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52d45420-b72b-650c-2b76-fb631180c51b" [ 1074.405246] env[61824]: _type = "Task" [ 1074.405246] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.413089] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d45420-b72b-650c-2b76-fb631180c51b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.544920] env[61824]: DEBUG oslo_concurrency.lockutils [None req-d56f6ef3-1d36-4986-9eb7-fa10d31c0736 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "a83ebb06-0343-4a84-a71f-a84e47f9ec5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.843s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.589800] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276129, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.917794] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52d45420-b72b-650c-2b76-fb631180c51b, 'name': SearchDatastore_Task, 'duration_secs': 0.078785} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.918655] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.918944] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.919277] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02107e93-8d69-44e1-8330-864729e520fc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.926509] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1074.926509] env[61824]: value = "task-1276130" [ 1074.926509] env[61824]: _type = "Task" [ 1074.926509] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.935592] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.070052] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.070411] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.103036] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276129, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.446571] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.580279] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.580950] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1075.581191] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Rebuilding the list of instances to heal {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1075.594525] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276129, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.943092] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276130, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.093543] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Skipping network cache update for instance because it is Building. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1076.095847] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276129, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.121629] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "4fc78f9e-91c5-4c28-8fa1-59f589747265" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.121629] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.137653] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.137831] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquired lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.137980] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Forcefully refreshing network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1076.138158] env[61824]: DEBUG nova.objects.instance [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lazy-loading 'info_cache' on Instance uuid 8aa00861-bca6-4861-a5e5-0c538155c9ab {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.438911] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276130, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.592325] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276129, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.181997} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.592559] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c263e67c-1047-4baa-835b-62b408c6dbe8/c263e67c-1047-4baa-835b-62b408c6dbe8.vmdk to [datastore2] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1076.593345] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada1bdc9-fe34-4e9b-9aa7-e2a1ffe4701c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.614632] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.614900] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f78203da-738f-4042-a63c-6aba384ccc09 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.628755] env[61824]: DEBUG nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1076.635975] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1076.635975] env[61824]: value = "task-1276131" [ 1076.635975] env[61824]: _type = "Task" [ 1076.635975] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.645894] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276131, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.938722] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276130, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.652966} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.939025] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1076.939268] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1076.939525] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-086f0b7f-989d-4db4-b057-37251858e314 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.945465] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1076.945465] env[61824]: value = "task-1276132" [ 1076.945465] env[61824]: _type = "Task" [ 1076.945465] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.954516] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276132, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.147662] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276131, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.254364] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.254634] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.256491] env[61824]: INFO nova.compute.claims [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1077.455038] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276132, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070732} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.455255] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1077.456031] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b2494a-3ba5-408b-980c-4f8350239e4d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.477394] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1077.477640] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5589b949-c726-4747-8e92-e71f50de9a14 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.496865] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1077.496865] env[61824]: value = "task-1276133" [ 1077.496865] env[61824]: _type = "Task" [ 1077.496865] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.504293] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.646276] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276131, 'name': ReconfigVM_Task, 'duration_secs': 0.534349} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.646653] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a/5bf9f3b5-7084-4c5a-952b-d90c337ad10a.vmdk or device None with type streamOptimized {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.647335] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92359a91-1340-45fe-a60d-3859b51c545c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.653789] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1077.653789] env[61824]: value = "task-1276134" [ 1077.653789] env[61824]: _type = "Task" [ 1077.653789] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.661571] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276134, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.853340] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [{"id": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "address": "fa:16:3e:83:f8:c1", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ec4b1c-0d", "ovs_interfaceid": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.006870] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276133, 'name': ReconfigVM_Task, 'duration_secs': 0.297692} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.007237] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.007714] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-474c1ff0-a25f-482c-a093-ba260f4e4d47 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.014225] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1078.014225] env[61824]: value = "task-1276135" [ 1078.014225] env[61824]: _type = "Task" [ 1078.014225] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.021693] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276135, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.163311] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276134, 'name': Rename_Task, 'duration_secs': 0.169791} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.163597] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1078.163842] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f086580c-827e-4bef-b8f4-b6854c55fe52 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.169384] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1078.169384] env[61824]: value = "task-1276136" [ 1078.169384] env[61824]: _type = "Task" [ 1078.169384] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.176762] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276136, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.351157] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b826201-607d-4195-a342-dae2473cca52 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.355354] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Releasing lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.355564] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updated the network info_cache for instance {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1078.358815] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.359073] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.360015] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6cfcbf-8fa6-4f92-971a-f2cf1ba16f85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.362953] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.363150] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.363648] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.363843] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.363983] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1078.364145] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.391164] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8903e65-4904-446c-ba98-804ab0f47409 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.398722] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a53c7ae-b508-4a66-a936-91b2dad98b34 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.412954] env[61824]: DEBUG nova.compute.provider_tree [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.523931] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276135, 'name': Rename_Task, 'duration_secs': 0.129372} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.524188] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1078.524424] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fba5bd4c-0fc0-4e90-9e5d-82cd1cac85a1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.530380] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1078.530380] env[61824]: value = "task-1276137" [ 1078.530380] env[61824]: _type = "Task" [ 1078.530380] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.537614] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276137, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.678315] env[61824]: DEBUG oslo_vmware.api [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276136, 'name': PowerOnVM_Task, 'duration_secs': 0.425439} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.678586] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.783748] env[61824]: DEBUG nova.compute.manager [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.784695] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8666a273-8e88-405b-bd16-64782e22bfb8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.867641] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.916834] env[61824]: DEBUG nova.scheduler.client.report [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.041285] env[61824]: DEBUG oslo_vmware.api [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276137, 'name': PowerOnVM_Task, 'duration_secs': 0.428533} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.041726] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1079.041784] env[61824]: INFO nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Took 10.35 seconds to spawn the instance on the hypervisor. [ 1079.041917] env[61824]: DEBUG nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1079.042701] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2d1d08-a141-4744-9a42-5253722be68c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.301540] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c117b97b-cba4-497e-b2f5-23067741a22b tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.131s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.423185] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.168s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.423731] env[61824]: DEBUG nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1079.426645] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.559s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.426828] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.426980] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1079.427837] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815f1e73-9ca3-4607-a502-d85bdcd668af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.437048] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867907f9-a915-407a-9dff-46862ae25220 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.450992] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b98d135-9049-4123-bc51-2015b776564b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.457008] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521b43fd-3b2a-44d4-a74e-cc1cc3ec0b5f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.485282] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180729MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1079.485434] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.485634] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.557726] env[61824]: INFO nova.compute.manager [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Took 15.99 seconds to build instance. [ 1079.932671] env[61824]: DEBUG nova.compute.utils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.934072] env[61824]: DEBUG nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1079.934241] env[61824]: DEBUG nova.network.neutron [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1079.981663] env[61824]: DEBUG nova.policy [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8763f64818e24fad80ce232ac7d4463f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '629276280a234e7a8b810e6e74779fff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1080.060748] env[61824]: DEBUG oslo_concurrency.lockutils [None req-40b8addf-ce1e-418f-b04a-a137876c79d3 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.501s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.229199] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "b8db876d-e430-4c5f-a439-0601dbf4598d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.229492] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.437757] env[61824]: DEBUG nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1080.453710] env[61824]: DEBUG nova.network.neutron [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Successfully created port: 47b8425d-b22b-414c-a46e-ccfb81f19895 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1080.510087] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8aa00861-bca6-4861-a5e5-0c538155c9ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.510313] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 4700de90-7230-43a5-87ae-f2c7ed9d4ea6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.510442] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 1ddb6d7a-d8e1-492d-84c9-d61e27faa105 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.510564] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance b8db876d-e430-4c5f-a439-0601dbf4598d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.510680] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 5bf9f3b5-7084-4c5a-952b-d90c337ad10a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.510793] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.510906] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 4fc78f9e-91c5-4c28-8fa1-59f589747265 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1080.511119] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1080.511297] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1080.604809] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ab0b05-5e00-46c2-aa4a-0aa50c03bc86 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.613282] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78424661-a79a-4077-b822-e4180e5742aa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.643357] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59389f33-5165-4ec0-899d-64a93340f8de {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.650631] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79e9580-f078-4282-ae90-231423762631 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.664895] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.699865] env[61824]: DEBUG nova.compute.manager [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Received event network-changed-7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.700068] env[61824]: DEBUG nova.compute.manager [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Refreshing instance network info cache due to event network-changed-7afd60b2-5996-4b14-9a2a-2c271dc1394c. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1080.700371] env[61824]: DEBUG oslo_concurrency.lockutils [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] Acquiring lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.700521] env[61824]: DEBUG oslo_concurrency.lockutils [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] Acquired lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.700690] env[61824]: DEBUG nova.network.neutron [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Refreshing network info cache for port 7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1080.732537] env[61824]: DEBUG nova.compute.utils [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1081.169092] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1081.235055] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.424161] env[61824]: DEBUG nova.network.neutron [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updated VIF entry in instance network info cache for port 7afd60b2-5996-4b14-9a2a-2c271dc1394c. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1081.424548] env[61824]: DEBUG nova.network.neutron [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updating instance_info_cache with network_info: [{"id": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "address": "fa:16:3e:35:ae:d2", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7afd60b2-59", "ovs_interfaceid": "7afd60b2-5996-4b14-9a2a-2c271dc1394c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.451974] env[61824]: DEBUG nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1081.477293] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1081.477697] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1081.477899] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1081.478141] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1081.478323] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1081.478500] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1081.478745] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1081.478938] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1081.479149] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1081.479339] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1081.479568] env[61824]: DEBUG nova.virt.hardware [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1081.480562] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f5285d-5d6e-4799-b009-6613b37322f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.488765] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95c5be9-ad6f-42a0-8ca7-5a17bbb9da80 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.674857] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1081.675041] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.189s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.675244] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1081.675384] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Cleaning up deleted instances {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1081.927799] env[61824]: DEBUG oslo_concurrency.lockutils [req-36ccd80d-f1dd-4ec3-9ca1-f645ce946358 req-6ca6606b-29cf-4c58-9e45-2bf5db78262c service nova] Releasing lock "refresh_cache-4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.080266] env[61824]: DEBUG nova.network.neutron [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Successfully updated port: 47b8425d-b22b-414c-a46e-ccfb81f19895 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1082.189838] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] There are 50 instances to clean {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1082.190208] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: a83ebb06-0343-4a84-a71f-a84e47f9ec5e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.305576] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "b8db876d-e430-4c5f-a439-0601dbf4598d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.305854] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.306109] env[61824]: INFO nova.compute.manager [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Attaching volume ccf8167d-4a22-469a-9a8b-c7baa195387f to /dev/sdb [ 1082.336279] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8bce95-a3ae-4add-a704-80d8cd6235b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.343074] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752056e4-cbe6-4b06-832b-5d86f8d6e224 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.356414] env[61824]: DEBUG nova.virt.block_device [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating existing volume attachment record: 7af027f2-703f-46d0-97bf-71781fe0a412 {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1082.582882] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-4fc78f9e-91c5-4c28-8fa1-59f589747265" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.582882] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-4fc78f9e-91c5-4c28-8fa1-59f589747265" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.583177] env[61824]: DEBUG nova.network.neutron [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1082.692999] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 358cd7ec-99b4-446c-9e60-2fe50d375893] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.726645] env[61824]: DEBUG nova.compute.manager [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Received event network-vif-plugged-47b8425d-b22b-414c-a46e-ccfb81f19895 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.726771] env[61824]: DEBUG oslo_concurrency.lockutils [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] Acquiring lock "4fc78f9e-91c5-4c28-8fa1-59f589747265-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.726994] env[61824]: DEBUG oslo_concurrency.lockutils [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.727202] env[61824]: DEBUG oslo_concurrency.lockutils [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.727391] env[61824]: DEBUG nova.compute.manager [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] No waiting events found dispatching network-vif-plugged-47b8425d-b22b-414c-a46e-ccfb81f19895 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1082.727581] env[61824]: WARNING nova.compute.manager [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Received unexpected event network-vif-plugged-47b8425d-b22b-414c-a46e-ccfb81f19895 for instance with vm_state building and task_state spawning. [ 1082.727760] env[61824]: DEBUG nova.compute.manager [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Received event network-changed-47b8425d-b22b-414c-a46e-ccfb81f19895 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.727927] env[61824]: DEBUG nova.compute.manager [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Refreshing instance network info cache due to event network-changed-47b8425d-b22b-414c-a46e-ccfb81f19895. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1082.728178] env[61824]: DEBUG oslo_concurrency.lockutils [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] Acquiring lock "refresh_cache-4fc78f9e-91c5-4c28-8fa1-59f589747265" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.125752] env[61824]: DEBUG nova.network.neutron [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1083.196037] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: a44b2d44-ecab-412a-b22e-c5a720b59631] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.267730] env[61824]: DEBUG nova.network.neutron [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Updating instance_info_cache with network_info: [{"id": "47b8425d-b22b-414c-a46e-ccfb81f19895", "address": "fa:16:3e:88:ac:26", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47b8425d-b2", "ovs_interfaceid": "47b8425d-b22b-414c-a46e-ccfb81f19895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.699383] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: d97da7c1-d20d-45db-a03d-c9d4ceb8c235] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.769739] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-4fc78f9e-91c5-4c28-8fa1-59f589747265" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.770072] env[61824]: DEBUG nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Instance network_info: |[{"id": "47b8425d-b22b-414c-a46e-ccfb81f19895", "address": "fa:16:3e:88:ac:26", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47b8425d-b2", "ovs_interfaceid": "47b8425d-b22b-414c-a46e-ccfb81f19895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1083.770433] env[61824]: DEBUG oslo_concurrency.lockutils [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] Acquired lock "refresh_cache-4fc78f9e-91c5-4c28-8fa1-59f589747265" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.770620] env[61824]: DEBUG nova.network.neutron [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Refreshing network info cache for port 47b8425d-b22b-414c-a46e-ccfb81f19895 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1083.772058] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:ac:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47b8425d-b22b-414c-a46e-ccfb81f19895', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1083.779432] env[61824]: DEBUG oslo.service.loopingcall [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1083.782187] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1083.782654] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69c2c7b2-c437-44d3-b9e2-becd93b6a85e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.804313] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1083.804313] env[61824]: value = "task-1276141" [ 1083.804313] env[61824]: _type = "Task" [ 1083.804313] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.812119] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276141, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.987901] env[61824]: DEBUG nova.network.neutron [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Updated VIF entry in instance network info cache for port 47b8425d-b22b-414c-a46e-ccfb81f19895. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1083.988352] env[61824]: DEBUG nova.network.neutron [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Updating instance_info_cache with network_info: [{"id": "47b8425d-b22b-414c-a46e-ccfb81f19895", "address": "fa:16:3e:88:ac:26", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47b8425d-b2", "ovs_interfaceid": "47b8425d-b22b-414c-a46e-ccfb81f19895", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.203187] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 1a0a00e9-c7ee-47c2-9c17-ee94c1fa116d] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.314436] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276141, 'name': CreateVM_Task, 'duration_secs': 0.291017} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.314597] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1084.315304] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.315489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.315804] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1084.316073] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cc7c928-15d3-4c00-aa08-14f25feba3ed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.320286] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1084.320286] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521fb7da-7025-43b0-04b6-c468d18b810b" [ 1084.320286] env[61824]: _type = "Task" [ 1084.320286] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.327373] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521fb7da-7025-43b0-04b6-c468d18b810b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.491019] env[61824]: DEBUG oslo_concurrency.lockutils [req-502f5091-2826-4b41-93d0-034debd7b203 req-7d32ff72-cb07-406d-9cea-b898b4336204 service nova] Releasing lock "refresh_cache-4fc78f9e-91c5-4c28-8fa1-59f589747265" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.706339] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: ccb21ac1-6793-4cec-9c7e-824e5aecb488] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.829979] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521fb7da-7025-43b0-04b6-c468d18b810b, 'name': SearchDatastore_Task, 'duration_secs': 0.009717} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.830313] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.830580] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1084.830822] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.830975] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.831177] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1084.831435] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1f18432-c3ac-4706-910a-49c9f8c10aca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.839275] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1084.839452] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1084.840130] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bf5ba41-655c-4694-a83b-fbe506d4648f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.845928] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1084.845928] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5231251f-919e-1955-a031-d077aef7cefe" [ 1084.845928] env[61824]: _type = "Task" [ 1084.845928] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.852906] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231251f-919e-1955-a031-d077aef7cefe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.210046] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: f29ed26b-94ed-43a5-baf0-84b6ba26fc22] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.357101] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5231251f-919e-1955-a031-d077aef7cefe, 'name': SearchDatastore_Task, 'duration_secs': 0.009159} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.357101] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e0b8745-8ed0-49a3-8293-3dadb01d20cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.362014] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1085.362014] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52f75e38-5e66-18ad-132c-bf7b75bbb812" [ 1085.362014] env[61824]: _type = "Task" [ 1085.362014] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.369656] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f75e38-5e66-18ad-132c-bf7b75bbb812, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.712667] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 82928174-1761-4296-bebe-7989c088d542] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.872042] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52f75e38-5e66-18ad-132c-bf7b75bbb812, 'name': SearchDatastore_Task, 'duration_secs': 0.00933} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.872263] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.872527] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4fc78f9e-91c5-4c28-8fa1-59f589747265/4fc78f9e-91c5-4c28-8fa1-59f589747265.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1085.872780] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efd7fe8a-692d-47b7-bf3e-7767b6d20c17 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.878636] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1085.878636] env[61824]: value = "task-1276143" [ 1085.878636] env[61824]: _type = "Task" [ 1085.878636] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.885823] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276143, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.216273] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 3256627d-35c8-4b75-b4a8-390eda0a150e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.388726] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276143, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471456} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.388988] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4fc78f9e-91c5-4c28-8fa1-59f589747265/4fc78f9e-91c5-4c28-8fa1-59f589747265.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1086.389238] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1086.389488] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae84e8f6-be95-429e-97db-a3b5da6b69f7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.395194] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1086.395194] env[61824]: value = "task-1276144" [ 1086.395194] env[61824]: _type = "Task" [ 1086.395194] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.401923] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.722941] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 44ec5c22-d080-45d1-aa21-080af4c317c5] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.900158] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1086.900439] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274245', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'name': 'volume-ccf8167d-4a22-469a-9a8b-c7baa195387f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8db876d-e430-4c5f-a439-0601dbf4598d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'serial': 'ccf8167d-4a22-469a-9a8b-c7baa195387f'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1086.901245] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f036229a-e944-4072-ae69-75fb1e7f3242 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.908779] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066866} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.918992] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.919799] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67065798-da2e-4ebd-9195-68f19c57624b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.922441] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcbd5b8-822d-4437-9ae8-f9b4f542947f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.945022] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] 4fc78f9e-91c5-4c28-8fa1-59f589747265/4fc78f9e-91c5-4c28-8fa1-59f589747265.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.957437] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dec369cd-649a-41b5-b344-b22a3ee87625 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.978125] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-ccf8167d-4a22-469a-9a8b-c7baa195387f/volume-ccf8167d-4a22-469a-9a8b-c7baa195387f.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.978390] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6d4efe9-9f09-426b-b7f6-8b32b2a98bff {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.996365] env[61824]: DEBUG oslo_vmware.api [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1086.996365] env[61824]: value = "task-1276145" [ 1086.996365] env[61824]: _type = "Task" [ 1086.996365] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.997546] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1086.997546] env[61824]: value = "task-1276146" [ 1086.997546] env[61824]: _type = "Task" [ 1086.997546] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.008324] env[61824]: DEBUG oslo_vmware.api [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276145, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.011296] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276146, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.226572] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e500f768-3ec8-48c7-a947-784e8665af15] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.510299] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276146, 'name': ReconfigVM_Task, 'duration_secs': 0.264733} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.513107] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Reconfigured VM instance instance-0000006c to attach disk [datastore1] 4fc78f9e-91c5-4c28-8fa1-59f589747265/4fc78f9e-91c5-4c28-8fa1-59f589747265.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.513796] env[61824]: DEBUG oslo_vmware.api [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276145, 'name': ReconfigVM_Task, 'duration_secs': 0.330548} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.514013] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f93b4f60-2e29-48ed-a6c1-3531469eb48b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.515449] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-ccf8167d-4a22-469a-9a8b-c7baa195387f/volume-ccf8167d-4a22-469a-9a8b-c7baa195387f.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.520099] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50d74f0a-bda5-45f9-8a97-465b1a377c39 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.534113] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1087.534113] env[61824]: value = "task-1276147" [ 1087.534113] env[61824]: _type = "Task" [ 1087.534113] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.535209] env[61824]: DEBUG oslo_vmware.api [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1087.535209] env[61824]: value = "task-1276148" [ 1087.535209] env[61824]: _type = "Task" [ 1087.535209] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.544951] env[61824]: DEBUG oslo_vmware.api [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.547807] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276147, 'name': Rename_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.730148] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 957ae086-5bb4-44f6-b287-5fbcd9742d12] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.052408] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276147, 'name': Rename_Task, 'duration_secs': 0.137001} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.055263] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1088.055537] env[61824]: DEBUG oslo_vmware.api [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276148, 'name': ReconfigVM_Task, 'duration_secs': 0.159594} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.055733] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d1dea57-0aab-4a72-a8b7-e38d7fe07701 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.057176] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274245', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'name': 'volume-ccf8167d-4a22-469a-9a8b-c7baa195387f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8db876d-e430-4c5f-a439-0601dbf4598d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'serial': 'ccf8167d-4a22-469a-9a8b-c7baa195387f'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1088.063231] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1088.063231] env[61824]: value = "task-1276149" [ 1088.063231] env[61824]: _type = "Task" [ 1088.063231] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.070096] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276149, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.233601] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: c5da84f7-4ee7-4537-a3f5-b95ee513c4dd] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.573945] env[61824]: DEBUG oslo_vmware.api [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276149, 'name': PowerOnVM_Task, 'duration_secs': 0.466167} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.574252] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.574459] env[61824]: INFO nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Took 7.12 seconds to spawn the instance on the hypervisor. [ 1088.574663] env[61824]: DEBUG nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.575471] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4987a154-6760-4c5a-ad0c-23fb06f24ad1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.736926] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 1f8c65c3-dd7a-4d70-a063-c3cf03b0b66e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.091610] env[61824]: INFO nova.compute.manager [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Took 11.96 seconds to build instance. [ 1089.093821] env[61824]: DEBUG nova.objects.instance [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lazy-loading 'flavor' on Instance uuid b8db876d-e430-4c5f-a439-0601dbf4598d {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.239756] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 16a84291-3bb9-49a6-9c24-ca399fd22e4e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.315825] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e585aee-26fc-4301-bcbe-b249aaba19af {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.322777] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f00bd5cb-f633-4cd5-a3eb-b4b1d817af35 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Suspending the VM {{(pid=61824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1089.323021] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ede0de05-916c-4dd3-8204-fb3efaa5d67e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.328907] env[61824]: DEBUG oslo_vmware.api [None req-f00bd5cb-f633-4cd5-a3eb-b4b1d817af35 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1089.328907] env[61824]: value = "task-1276150" [ 1089.328907] env[61824]: _type = "Task" [ 1089.328907] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.336810] env[61824]: DEBUG oslo_vmware.api [None req-f00bd5cb-f633-4cd5-a3eb-b4b1d817af35 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276150, 'name': SuspendVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.594107] env[61824]: DEBUG oslo_concurrency.lockutils [None req-94feb9a3-e76e-4409-9da7-2229eead2a59 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.472s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.599194] env[61824]: DEBUG oslo_concurrency.lockutils [None req-32400594-994d-4b32-990d-9af05a523e27 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.293s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.664752] env[61824]: INFO nova.compute.manager [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Rescuing [ 1089.665081] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.665259] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.665417] env[61824]: DEBUG nova.network.neutron [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1089.742860] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e5825315-bfea-4eec-80ca-cb68d9748d22] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.838159] env[61824]: DEBUG oslo_vmware.api [None req-f00bd5cb-f633-4cd5-a3eb-b4b1d817af35 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276150, 'name': SuspendVM_Task} progress is 70%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.246379] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: beafe3e0-ff41-40f7-90cf-1d7977191180] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.340136] env[61824]: DEBUG oslo_vmware.api [None req-f00bd5cb-f633-4cd5-a3eb-b4b1d817af35 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276150, 'name': SuspendVM_Task, 'duration_secs': 0.648147} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.340789] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-f00bd5cb-f633-4cd5-a3eb-b4b1d817af35 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Suspended the VM {{(pid=61824) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1090.340980] env[61824]: DEBUG nova.compute.manager [None req-f00bd5cb-f633-4cd5-a3eb-b4b1d817af35 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.341827] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c588eb-9cc2-42d0-8038-e62523543374 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.439736] env[61824]: DEBUG nova.network.neutron [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.749443] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 0bb7e086-dcdd-4a5e-a91a-d1fa183fcde8] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.943114] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.252503] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 07973a77-0fe8-4bc9-b8c5-e693d7a63ef5] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.474606] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1091.474880] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7748f6a1-d227-4d04-a213-16bf3f38fd33 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.481994] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1091.481994] env[61824]: value = "task-1276151" [ 1091.481994] env[61824]: _type = "Task" [ 1091.481994] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.489844] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276151, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.708534] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "4fc78f9e-91c5-4c28-8fa1-59f589747265" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.708737] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.708949] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "4fc78f9e-91c5-4c28-8fa1-59f589747265-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.709155] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.709339] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.711625] env[61824]: INFO nova.compute.manager [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Terminating instance [ 1091.713519] env[61824]: DEBUG nova.compute.manager [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1091.713722] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1091.714604] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cc9bc9-a37e-4dae-a49a-85e3d92ef7b0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.721717] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1091.721954] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb02b69d-7118-476d-b24f-43bc96149325 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.755560] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 9f031313-766b-4696-9bc6-75ced078bfde] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.784710] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1091.785096] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1091.785318] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleting the datastore file [datastore1] 4fc78f9e-91c5-4c28-8fa1-59f589747265 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1091.785582] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd1c4c25-6549-40d5-b726-25595ba2d076 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.792106] env[61824]: DEBUG oslo_vmware.api [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1091.792106] env[61824]: value = "task-1276153" [ 1091.792106] env[61824]: _type = "Task" [ 1091.792106] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.799549] env[61824]: DEBUG oslo_vmware.api [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276153, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.994719] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276151, 'name': PowerOffVM_Task, 'duration_secs': 0.190092} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.995105] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.996321] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed78a44e-d060-4f28-a858-65652f0e46ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.028219] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d035f8-dcb3-4864-ab49-af3d3467ddb1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.058020] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1092.058414] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e0e061e6-4a69-4606-a5f7-c9d74a6356c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.065075] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1092.065075] env[61824]: value = "task-1276154" [ 1092.065075] env[61824]: _type = "Task" [ 1092.065075] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.071915] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.258636] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 09432e72-c10a-4a7a-93af-af0d371fb871] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.301461] env[61824]: DEBUG oslo_vmware.api [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276153, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140023} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.301716] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1092.301903] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1092.302093] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1092.302271] env[61824]: INFO nova.compute.manager [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1092.302526] env[61824]: DEBUG oslo.service.loopingcall [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.302732] env[61824]: DEBUG nova.compute.manager [-] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1092.302826] env[61824]: DEBUG nova.network.neutron [-] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1092.575753] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1092.575979] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1092.576273] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.576403] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.576587] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1092.576841] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d181245-334f-401e-b086-33c48d8e7c95 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.587904] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1092.588170] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1092.588927] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ceea995-d636-46da-bd98-933d0e2da477 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.594102] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1092.594102] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5267e4b7-413a-a975-c31b-a9a30e910616" [ 1092.594102] env[61824]: _type = "Task" [ 1092.594102] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.601229] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5267e4b7-413a-a975-c31b-a9a30e910616, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.668895] env[61824]: DEBUG nova.compute.manager [req-1889640a-9c34-4b15-9b47-1057af9365d9 req-2063cc2c-da10-4ff1-b3fb-883a67b32296 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Received event network-vif-deleted-47b8425d-b22b-414c-a46e-ccfb81f19895 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.669100] env[61824]: INFO nova.compute.manager [req-1889640a-9c34-4b15-9b47-1057af9365d9 req-2063cc2c-da10-4ff1-b3fb-883a67b32296 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Neutron deleted interface 47b8425d-b22b-414c-a46e-ccfb81f19895; detaching it from the instance and deleting it from the info cache [ 1092.669283] env[61824]: DEBUG nova.network.neutron [req-1889640a-9c34-4b15-9b47-1057af9365d9 req-2063cc2c-da10-4ff1-b3fb-883a67b32296 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.762140] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: b1a770f9-71ea-4f80-a562-c08a313b5753] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.104426] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5267e4b7-413a-a975-c31b-a9a30e910616, 'name': SearchDatastore_Task, 'duration_secs': 0.008468} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.105180] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fd1dca6-ea66-438d-9b61-7f6399b91219 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.109974] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1093.109974] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52073082-ef1f-941e-1922-9f4cd23b8183" [ 1093.109974] env[61824]: _type = "Task" [ 1093.109974] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.117101] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52073082-ef1f-941e-1922-9f4cd23b8183, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.142479] env[61824]: DEBUG nova.network.neutron [-] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.172609] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dca5254-a99a-4384-9f66-37913b10d52e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.182495] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5112ad8c-6ad2-4812-9cdf-8b9b4995408f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.209414] env[61824]: DEBUG nova.compute.manager [req-1889640a-9c34-4b15-9b47-1057af9365d9 req-2063cc2c-da10-4ff1-b3fb-883a67b32296 service nova] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Detach interface failed, port_id=47b8425d-b22b-414c-a46e-ccfb81f19895, reason: Instance 4fc78f9e-91c5-4c28-8fa1-59f589747265 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1093.265838] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: c6552176-1128-473b-ac4a-f3b55428fd9e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.620725] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52073082-ef1f-941e-1922-9f4cd23b8183, 'name': SearchDatastore_Task, 'duration_secs': 0.008077} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.620999] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.621291] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk. {{(pid=61824) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1093.621573] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f7f834a-f9f1-47aa-ba1e-752fa0d36fbe {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.627367] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1093.627367] env[61824]: value = "task-1276155" [ 1093.627367] env[61824]: _type = "Task" [ 1093.627367] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.634493] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276155, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.645011] env[61824]: INFO nova.compute.manager [-] [instance: 4fc78f9e-91c5-4c28-8fa1-59f589747265] Took 1.34 seconds to deallocate network for instance. [ 1093.768746] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: cabad7a6-206f-4e02-be3c-e18654396b7e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.136950] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276155, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448855} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.139287] env[61824]: INFO nova.virt.vmwareapi.ds_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk. [ 1094.139287] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cae5129-15d1-434d-994d-215ae9dca650 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.158052] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.158296] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.158518] env[61824]: DEBUG nova.objects.instance [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'resources' on Instance uuid 4fc78f9e-91c5-4c28-8fa1-59f589747265 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.166550] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1094.169460] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a21042d1-a91b-475e-8286-740ebeb9c453 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.186958] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1094.186958] env[61824]: value = "task-1276156" [ 1094.186958] env[61824]: _type = "Task" [ 1094.186958] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.194585] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276156, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.271768] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f45c359-b22e-4bd3-bc8b-bc5120f752f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.274449] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 5806faad-2bab-4fda-a131-566af50c4d7e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.280918] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c879f452-dc67-49d2-8365-1b0a0928ed3d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.311621] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3dead84-bb18-4260-9ef3-2b0fd2be1c9e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.318713] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967daeac-9068-45e7-a895-bd04bd70117b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.331770] env[61824]: DEBUG nova.compute.provider_tree [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1094.696028] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276156, 'name': ReconfigVM_Task, 'duration_secs': 0.296228} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.696300] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d/11b2670a-52f3-4a0e-8363-54ca7a7f5264-rescue.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.697167] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3b24d5-6ec5-42ec-9d78-eb9428691321 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.723820] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e7b8df9-b8fd-4d5d-811c-51d9f5f12580 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.738791] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1094.738791] env[61824]: value = "task-1276157" [ 1094.738791] env[61824]: _type = "Task" [ 1094.738791] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.746230] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276157, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.777835] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: aa9c9eb2-b411-4eca-9f8d-3df377ffe351] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.834620] env[61824]: DEBUG nova.scheduler.client.report [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.250023] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276157, 'name': ReconfigVM_Task, 'duration_secs': 0.148768} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.250319] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1095.250580] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66afdce1-729c-4aa1-a7ae-6bf19c12f65b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.257643] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1095.257643] env[61824]: value = "task-1276158" [ 1095.257643] env[61824]: _type = "Task" [ 1095.257643] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.265068] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.280584] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 37c405be-62a1-4feb-8cae-ea1c924f4f40] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.339049] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.181s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.359518] env[61824]: INFO nova.scheduler.client.report [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted allocations for instance 4fc78f9e-91c5-4c28-8fa1-59f589747265 [ 1095.767703] env[61824]: DEBUG oslo_vmware.api [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276158, 'name': PowerOnVM_Task, 'duration_secs': 0.384374} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.768994] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.770911] env[61824]: DEBUG nova.compute.manager [None req-0a7da666-1726-42b7-bf6d-6316776cca4a tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1095.771687] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98aaba6-cc8b-41e6-aeae-a5e5398e0089 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.783510] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 31edad40-a743-4b65-8ef0-615a9ba2a225] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.868931] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8838fc22-d7b8-47ba-ac3b-16f881af1cc5 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "4fc78f9e-91c5-4c28-8fa1-59f589747265" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.160s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.285742] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 01fd9faa-a0de-45b2-89bf-094c6d1ac155] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.566022] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "89fa4d95-b305-4490-906d-903e5823b6de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.566022] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.788211] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 21759f11-80e9-419d-9f65-ca4767d1593b] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.951078] env[61824]: INFO nova.compute.manager [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Unrescuing [ 1096.951426] env[61824]: DEBUG oslo_concurrency.lockutils [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.951606] env[61824]: DEBUG oslo_concurrency.lockutils [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquired lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.951778] env[61824]: DEBUG nova.network.neutron [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1097.067778] env[61824]: DEBUG nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1097.291584] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 1f6fbd3e-df72-4fc0-b91f-97ca6d5dd576] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.589807] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.590087] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.591627] env[61824]: INFO nova.compute.claims [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.711673] env[61824]: DEBUG nova.network.neutron [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.794965] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 5569eb0a-fd42-4e2d-b805-bb40e3264a65] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.214535] env[61824]: DEBUG oslo_concurrency.lockutils [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Releasing lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.215196] env[61824]: DEBUG nova.objects.instance [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lazy-loading 'flavor' on Instance uuid b8db876d-e430-4c5f-a439-0601dbf4598d {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.297860] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: d6576f45-35f6-48dd-9160-80fc53d8b6b2] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1098.680868] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d123ad3-0a9d-4fe8-ac03-e4099307c08b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.688496] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff08734a-f3f2-4077-85c4-8f1b7cf4623c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.717415] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47e46fa-72b1-4de3-a911-17e5ea170787 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.722789] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c030587a-5ebf-4541-8de4-222c7e6d260f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.729180] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6729d186-ac20-437c-8f8c-0df0a03f78c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.750227] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1098.750858] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ee26c2e-d403-4bae-900e-44e9f0503559 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.763086] env[61824]: DEBUG nova.compute.provider_tree [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.765223] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1098.765223] env[61824]: value = "task-1276159" [ 1098.765223] env[61824]: _type = "Task" [ 1098.765223] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.772511] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.800961] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 2495fcb4-28cb-49c3-90fe-c84072466287] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.267406] env[61824]: DEBUG nova.scheduler.client.report [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1099.279317] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276159, 'name': PowerOffVM_Task, 'duration_secs': 0.214543} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.279567] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1099.284858] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfiguring VM instance instance-00000068 to detach disk 2002 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1099.285642] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb392099-2cee-4f77-9822-a1818cebb67e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.303143] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1099.303143] env[61824]: value = "task-1276160" [ 1099.303143] env[61824]: _type = "Task" [ 1099.303143] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.303531] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: cbded637-56b9-4048-b66e-1edf6311f435] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.312441] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276160, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.775979] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.186s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.776614] env[61824]: DEBUG nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1099.808806] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 43071be0-eaab-415c-8dcb-948bd9702885] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1099.813941] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276160, 'name': ReconfigVM_Task, 'duration_secs': 0.236402} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.814378] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfigured VM instance instance-00000068 to detach disk 2002 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1099.814578] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.815060] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2dfb67a4-3e4d-4d59-b0d1-435e433ed614 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.820712] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1099.820712] env[61824]: value = "task-1276161" [ 1099.820712] env[61824]: _type = "Task" [ 1099.820712] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.827793] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276161, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.281737] env[61824]: DEBUG nova.compute.utils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1100.282971] env[61824]: DEBUG nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1100.283153] env[61824]: DEBUG nova.network.neutron [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1100.315578] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7be79422-c624-4b4a-884f-29b13b9d81d2] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1100.319376] env[61824]: DEBUG nova.policy [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8763f64818e24fad80ce232ac7d4463f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '629276280a234e7a8b810e6e74779fff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1100.329153] env[61824]: DEBUG oslo_vmware.api [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276161, 'name': PowerOnVM_Task, 'duration_secs': 0.354138} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.329404] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1100.329768] env[61824]: DEBUG nova.compute.manager [None req-91c27141-8153-47a7-a95f-838086d9675b tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1100.330386] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393f8010-c943-4801-8b5a-fad499920c01 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.547194] env[61824]: DEBUG nova.network.neutron [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Successfully created port: 76ef6f05-28f7-4cff-bf0d-74bc13fdc771 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1100.786119] env[61824]: DEBUG nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1100.818599] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: d9bb3172-c058-4e59-a719-d8aa340fba71] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.321694] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 0201e783-ed75-4cee-8df4-efac97b4ca19] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.754466] env[61824]: DEBUG nova.compute.manager [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Received event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1101.754751] env[61824]: DEBUG nova.compute.manager [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing instance network info cache due to event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1101.754988] env[61824]: DEBUG oslo_concurrency.lockutils [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] Acquiring lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.755154] env[61824]: DEBUG oslo_concurrency.lockutils [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] Acquired lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.755325] env[61824]: DEBUG nova.network.neutron [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1101.795086] env[61824]: DEBUG nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1101.821025] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.821391] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.821548] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.821738] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.821950] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.822049] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.822257] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.822464] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.822593] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.822761] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.822941] env[61824]: DEBUG nova.virt.hardware [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.824151] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6c9e43-5526-4249-b471-61075e1c3319 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.827811] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 6235a9b9-19c0-4917-938a-bca82a2ca05e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1101.834889] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da67e42-df07-4633-a360-c064506ad97c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.993527] env[61824]: DEBUG nova.network.neutron [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Successfully updated port: 76ef6f05-28f7-4cff-bf0d-74bc13fdc771 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1102.330607] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8522fe80-4d6d-4474-ae63-00daa0dc7349] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.462317] env[61824]: DEBUG nova.network.neutron [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updated VIF entry in instance network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1102.462713] env[61824]: DEBUG nova.network.neutron [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.496692] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.496692] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.496692] env[61824]: DEBUG nova.network.neutron [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1102.833749] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: a33e2573-6d2f-4d07-89d4-f1c6d9d9b382] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1102.965917] env[61824]: DEBUG oslo_concurrency.lockutils [req-cce0cbf6-36b5-4817-b96e-290c61c015ed req-2d078380-c9e7-49bf-9c95-08d7b78d2ddd service nova] Releasing lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.034603] env[61824]: DEBUG nova.network.neutron [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1103.191703] env[61824]: DEBUG nova.network.neutron [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance_info_cache with network_info: [{"id": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "address": "fa:16:3e:d0:1e:20", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76ef6f05-28", "ovs_interfaceid": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.336503] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: e3e19a04-a29b-4473-833c-d8486703e9ad] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1103.693969] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.694330] env[61824]: DEBUG nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Instance network_info: |[{"id": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "address": "fa:16:3e:d0:1e:20", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76ef6f05-28", "ovs_interfaceid": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1103.694780] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:1e:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76ef6f05-28f7-4cff-bf0d-74bc13fdc771', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.702415] env[61824]: DEBUG oslo.service.loopingcall [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.702616] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.702841] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a97ba2e-09d9-4515-aba9-7f50cfab7db5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.722203] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.722203] env[61824]: value = "task-1276162" [ 1103.722203] env[61824]: _type = "Task" [ 1103.722203] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.729489] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276162, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.787227] env[61824]: DEBUG nova.compute.manager [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Received event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1103.787494] env[61824]: DEBUG nova.compute.manager [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing instance network info cache due to event network-changed-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1103.787678] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Acquiring lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.787854] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Acquired lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.787967] env[61824]: DEBUG nova.network.neutron [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Refreshing network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1103.839433] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: f24881d3-a647-42f9-bf82-0065f3425a5f] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.231584] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276162, 'name': CreateVM_Task, 'duration_secs': 0.2945} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.231798] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1104.232484] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.232660] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.233009] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1104.233272] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41243ae9-f5a5-48d2-8212-1a0566f17c92 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.237807] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1104.237807] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52abd425-6f8a-8b4c-b092-1e9c61849a34" [ 1104.237807] env[61824]: _type = "Task" [ 1104.237807] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.244845] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52abd425-6f8a-8b4c-b092-1e9c61849a34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.343308] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: cc13d6a5-6b99-40bf-8978-09d284b4ed17] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.486984] env[61824]: DEBUG nova.network.neutron [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updated VIF entry in instance network info cache for port c75f0bf1-26d7-46fd-a861-f3ddbf753d2d. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1104.487597] env[61824]: DEBUG nova.network.neutron [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [{"id": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "address": "fa:16:3e:9a:1a:bc", "network": {"id": "69db7da2-b4af-4adb-a2ba-c07b0ddfe0f0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-824861153-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f9aff6d76e534482b3c4d98034a9591a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b399c74-1411-408a-b4cd-84e268ae83fe", "external-id": "nsx-vlan-transportzone-486", "segmentation_id": 486, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc75f0bf1-26", "ovs_interfaceid": "c75f0bf1-26d7-46fd-a861-f3ddbf753d2d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.747777] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52abd425-6f8a-8b4c-b092-1e9c61849a34, 'name': SearchDatastore_Task, 'duration_secs': 0.011121} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.748071] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.748351] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.748585] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.748738] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.748920] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.749189] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f351c009-5932-4026-99a4-64464259ba58 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.756991] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.757185] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.757864] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98eea52a-4967-4703-a068-8b825ddda053 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.762681] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1104.762681] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]520e4557-5f37-d675-9913-f9371ab5a9e9" [ 1104.762681] env[61824]: _type = "Task" [ 1104.762681] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.770718] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520e4557-5f37-d675-9913-f9371ab5a9e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.847913] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 23fbfb1e-a716-4a73-a15d-16ff25690e0e] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1104.991102] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Releasing lock "refresh_cache-b8db876d-e430-4c5f-a439-0601dbf4598d" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.991416] env[61824]: DEBUG nova.compute.manager [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Received event network-vif-plugged-76ef6f05-28f7-4cff-bf0d-74bc13fdc771 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.991623] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Acquiring lock "89fa4d95-b305-4490-906d-903e5823b6de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.991838] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Lock "89fa4d95-b305-4490-906d-903e5823b6de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.992010] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Lock "89fa4d95-b305-4490-906d-903e5823b6de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.992193] env[61824]: DEBUG nova.compute.manager [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] No waiting events found dispatching network-vif-plugged-76ef6f05-28f7-4cff-bf0d-74bc13fdc771 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1104.992367] env[61824]: WARNING nova.compute.manager [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Received unexpected event network-vif-plugged-76ef6f05-28f7-4cff-bf0d-74bc13fdc771 for instance with vm_state building and task_state spawning. [ 1104.992530] env[61824]: DEBUG nova.compute.manager [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Received event network-changed-76ef6f05-28f7-4cff-bf0d-74bc13fdc771 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.992689] env[61824]: DEBUG nova.compute.manager [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Refreshing instance network info cache due to event network-changed-76ef6f05-28f7-4cff-bf0d-74bc13fdc771. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1104.992879] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Acquiring lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.993027] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Acquired lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.993191] env[61824]: DEBUG nova.network.neutron [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Refreshing network info cache for port 76ef6f05-28f7-4cff-bf0d-74bc13fdc771 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1105.272716] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]520e4557-5f37-d675-9913-f9371ab5a9e9, 'name': SearchDatastore_Task, 'duration_secs': 0.008037} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.273535] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d81b9f6-9af3-466c-9f31-c065eba3c69b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.278286] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1105.278286] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]5251fbb2-3126-89b7-18db-49f7142690b8" [ 1105.278286] env[61824]: _type = "Task" [ 1105.278286] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.285390] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5251fbb2-3126-89b7-18db-49f7142690b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.349939] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7f6a4446-86e1-44c7-ab42-297b033dbace] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1105.671742] env[61824]: DEBUG nova.network.neutron [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updated VIF entry in instance network info cache for port 76ef6f05-28f7-4cff-bf0d-74bc13fdc771. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1105.672107] env[61824]: DEBUG nova.network.neutron [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance_info_cache with network_info: [{"id": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "address": "fa:16:3e:d0:1e:20", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76ef6f05-28", "ovs_interfaceid": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.788325] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]5251fbb2-3126-89b7-18db-49f7142690b8, 'name': SearchDatastore_Task, 'duration_secs': 0.008536} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.788577] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.788835] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 89fa4d95-b305-4490-906d-903e5823b6de/89fa4d95-b305-4490-906d-903e5823b6de.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.789113] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5c5d36c-2712-4ee7-bb24-e937d272113b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.795141] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1105.795141] env[61824]: value = "task-1276163" [ 1105.795141] env[61824]: _type = "Task" [ 1105.795141] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.802614] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.853390] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 7e73f8b5-a138-4455-a392-9a2b2b860558] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.175029] env[61824]: DEBUG oslo_concurrency.lockutils [req-5692ccac-b2de-4a3e-905d-51d4ef38df32 req-62e0acc8-2d14-422e-9c1f-179c427296b2 service nova] Releasing lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.305203] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276163, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439311} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.305475] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 89fa4d95-b305-4490-906d-903e5823b6de/89fa4d95-b305-4490-906d-903e5823b6de.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1106.305747] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1106.306035] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b9260d36-2cf0-455e-ab17-094fce587e99 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.312575] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1106.312575] env[61824]: value = "task-1276164" [ 1106.312575] env[61824]: _type = "Task" [ 1106.312575] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.319791] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276164, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.356501] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 9fd8e236-3556-4b95-952f-3c324b896e29] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.822672] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276164, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057088} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.822935] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1106.823689] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b97cbb-7a48-40af-8805-dcecc4e753b1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.844692] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 89fa4d95-b305-4490-906d-903e5823b6de/89fa4d95-b305-4490-906d-903e5823b6de.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.844925] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3ef1b7f8-63ff-4a45-9117-8322b4782f78 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.859676] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 3f89fb2d-99c1-4cf7-83aa-db033211a2ce] Instance has had 0 of 5 cleanup attempts {{(pid=61824) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1106.866340] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1106.866340] env[61824]: value = "task-1276165" [ 1106.866340] env[61824]: _type = "Task" [ 1106.866340] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.873573] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.363506] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.363846] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Cleaning up deleted instances with incomplete migration {{(pid=61824) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1107.377154] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.866950] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1107.876780] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276165, 'name': ReconfigVM_Task, 'duration_secs': 0.716568} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.877044] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 89fa4d95-b305-4490-906d-903e5823b6de/89fa4d95-b305-4490-906d-903e5823b6de.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.878139] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58bab917-5d7a-40e7-b4c3-b082506385fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.884105] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1107.884105] env[61824]: value = "task-1276166" [ 1107.884105] env[61824]: _type = "Task" [ 1107.884105] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.891702] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276166, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.394538] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276166, 'name': Rename_Task, 'duration_secs': 0.121297} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.395286] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1108.395543] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe97b1d8-3c46-4979-ac00-8fb3e320e29d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.401712] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1108.401712] env[61824]: value = "task-1276167" [ 1108.401712] env[61824]: _type = "Task" [ 1108.401712] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.408976] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276167, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.912191] env[61824]: DEBUG oslo_vmware.api [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276167, 'name': PowerOnVM_Task, 'duration_secs': 0.456078} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.912507] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.912683] env[61824]: INFO nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Took 7.12 seconds to spawn the instance on the hypervisor. [ 1108.913051] env[61824]: DEBUG nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1108.913659] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d733e4-c6ec-41c1-8aa5-7aa418de2cea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.431501] env[61824]: INFO nova.compute.manager [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Took 11.86 seconds to build instance. [ 1109.933898] env[61824]: DEBUG oslo_concurrency.lockutils [None req-5c4a8039-cb24-4854-b5e3-6160692fb8ca tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.368s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.706025] env[61824]: DEBUG nova.compute.manager [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Stashing vm_state: active {{(pid=61824) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1111.225671] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.225936] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.731208] env[61824]: INFO nova.compute.claims [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1112.237542] env[61824]: INFO nova.compute.resource_tracker [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating resource usage from migration 559c6e8d-7c7e-4560-a123-167e6b42318f [ 1112.330303] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63dede52-e2c9-4cca-9be8-aee4f0149336 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.337796] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db54bad0-ad7e-4f42-862b-5e7fa6f914fb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.366096] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5456c7-6e5e-454c-901e-625cd4c788f4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.372894] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4b3a0f-f3b0-4709-a899-dd02dc09cf39 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.385472] env[61824]: DEBUG nova.compute.provider_tree [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1112.888433] env[61824]: DEBUG nova.scheduler.client.report [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.393432] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.167s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.393631] env[61824]: INFO nova.compute.manager [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Migrating [ 1113.909366] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.909734] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.909829] env[61824]: DEBUG nova.network.neutron [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1114.601943] env[61824]: DEBUG nova.network.neutron [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance_info_cache with network_info: [{"id": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "address": "fa:16:3e:d0:1e:20", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76ef6f05-28", "ovs_interfaceid": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.105052] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1116.172209] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.172559] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.172666] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.172885] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.173042] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.175098] env[61824]: INFO nova.compute.manager [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Terminating instance [ 1116.176778] env[61824]: DEBUG nova.compute.manager [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1116.176970] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1116.177835] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b0009d-2074-4332-bfed-7838ec5ac572 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.185137] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1116.185386] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2239d539-a24a-4299-a817-75ce2780ccc2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.192025] env[61824]: DEBUG oslo_vmware.api [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1116.192025] env[61824]: value = "task-1276168" [ 1116.192025] env[61824]: _type = "Task" [ 1116.192025] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.199289] env[61824]: DEBUG oslo_vmware.api [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.619841] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0eab7a9-1c17-4d88-bd41-b553ad52730e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.640151] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance '89fa4d95-b305-4490-906d-903e5823b6de' progress to 0 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1116.702242] env[61824]: DEBUG oslo_vmware.api [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276168, 'name': PowerOffVM_Task, 'duration_secs': 0.210892} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.702530] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1116.702708] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1116.702991] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e060e00-4e76-4536-935f-44b6af431cf5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.717422] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.717652] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.764103] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1116.764103] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1116.764103] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleting the datastore file [datastore2] 5bf9f3b5-7084-4c5a-952b-d90c337ad10a {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.764330] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f80633c-fbf8-4e7d-8932-b352f7669004 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.770623] env[61824]: DEBUG oslo_vmware.api [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for the task: (returnval){ [ 1116.770623] env[61824]: value = "task-1276170" [ 1116.770623] env[61824]: _type = "Task" [ 1116.770623] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.777969] env[61824]: DEBUG oslo_vmware.api [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.145790] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1117.146132] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-134a5fc8-db0a-40ea-9282-972e18047a39 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.152539] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1117.152539] env[61824]: value = "task-1276171" [ 1117.152539] env[61824]: _type = "Task" [ 1117.152539] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.159898] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276171, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.222244] env[61824]: DEBUG nova.compute.utils [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1117.280640] env[61824]: DEBUG oslo_vmware.api [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Task: {'id': task-1276170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11902} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.280903] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1117.281109] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1117.281299] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1117.281479] env[61824]: INFO nova.compute.manager [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1117.281734] env[61824]: DEBUG oslo.service.loopingcall [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.281938] env[61824]: DEBUG nova.compute.manager [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1117.282046] env[61824]: DEBUG nova.network.neutron [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1117.662471] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276171, 'name': PowerOffVM_Task, 'duration_secs': 0.196726} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.662676] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1117.662863] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance '89fa4d95-b305-4490-906d-903e5823b6de' progress to 17 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1117.724865] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.779236] env[61824]: DEBUG nova.compute.manager [req-fa012d9a-7209-4b45-bfaf-3faf53b310f7 req-2a8f5723-c1eb-4e63-9143-0a93aa85a60d service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Received event network-vif-deleted-4c770083-4d53-46bb-b9a4-cdaa981b457c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1117.779236] env[61824]: INFO nova.compute.manager [req-fa012d9a-7209-4b45-bfaf-3faf53b310f7 req-2a8f5723-c1eb-4e63-9143-0a93aa85a60d service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Neutron deleted interface 4c770083-4d53-46bb-b9a4-cdaa981b457c; detaching it from the instance and deleting it from the info cache [ 1117.779236] env[61824]: DEBUG nova.network.neutron [req-fa012d9a-7209-4b45-bfaf-3faf53b310f7 req-2a8f5723-c1eb-4e63-9143-0a93aa85a60d service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.169422] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1118.169605] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1118.169708] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1118.169898] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1118.170063] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1118.170235] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1118.170454] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1118.170656] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1118.170837] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1118.171013] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1118.171204] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1118.176271] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a872c49-994c-4856-a1b3-d1354fcf160d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.192276] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1118.192276] env[61824]: value = "task-1276172" [ 1118.192276] env[61824]: _type = "Task" [ 1118.192276] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.200047] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276172, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.258617] env[61824]: DEBUG nova.network.neutron [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.282264] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67b6da75-a0bc-4d26-98a7-038479445a8f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.297815] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4291a89-e34e-4f8b-977a-70af8d65fc2f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.324718] env[61824]: DEBUG nova.compute.manager [req-fa012d9a-7209-4b45-bfaf-3faf53b310f7 req-2a8f5723-c1eb-4e63-9143-0a93aa85a60d service nova] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Detach interface failed, port_id=4c770083-4d53-46bb-b9a4-cdaa981b457c, reason: Instance 5bf9f3b5-7084-4c5a-952b-d90c337ad10a could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1118.702499] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276172, 'name': ReconfigVM_Task, 'duration_secs': 0.273729} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.702811] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance '89fa4d95-b305-4490-906d-903e5823b6de' progress to 33 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1118.761035] env[61824]: INFO nova.compute.manager [-] [instance: 5bf9f3b5-7084-4c5a-952b-d90c337ad10a] Took 1.48 seconds to deallocate network for instance. [ 1118.793466] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.793719] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1118.793953] env[61824]: INFO nova.compute.manager [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Attaching volume bb2fd20c-89c6-4c6f-8752-040cc2888bf5 to /dev/sdb [ 1118.824495] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923754ae-4b76-477e-888f-9070af91d6ac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.831733] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c339e9e3-6253-4cf2-a782-fc0fe5e0f6b3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.843902] env[61824]: DEBUG nova.virt.block_device [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updating existing volume attachment record: 0125c4eb-f125-4143-9388-9de1647a69a1 {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1119.209532] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1119.209793] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1119.209991] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1119.210252] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1119.210442] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1119.210661] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1119.210912] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1119.211135] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1119.211340] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1119.211558] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1119.211756] env[61824]: DEBUG nova.virt.hardware [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1119.217420] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Reconfiguring VM instance instance-0000006d to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1119.217712] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1faef4f-3c96-4b5a-a0f6-368d10be2a43 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.235183] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1119.235183] env[61824]: value = "task-1276176" [ 1119.235183] env[61824]: _type = "Task" [ 1119.235183] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.242992] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276176, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.267138] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.267444] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.267620] env[61824]: DEBUG nova.objects.instance [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lazy-loading 'resources' on Instance uuid 5bf9f3b5-7084-4c5a-952b-d90c337ad10a {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.746046] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276176, 'name': ReconfigVM_Task, 'duration_secs': 0.145494} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.746046] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Reconfigured VM instance instance-0000006d to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1119.746530] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023d00d5-a8e0-406c-ba9b-76261634ade2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.769215] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 89fa4d95-b305-4490-906d-903e5823b6de/89fa4d95-b305-4490-906d-903e5823b6de.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1119.771516] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e53f411-956d-4295-afbd-0cceae781f8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.788407] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1119.788407] env[61824]: value = "task-1276177" [ 1119.788407] env[61824]: _type = "Task" [ 1119.788407] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.796364] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.885503] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2aae963-7f8a-46f1-8230-f0edd5d893cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.892768] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9b7aa9-44c1-420e-82ec-bcd28610f270 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.921915] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e890209-ab5a-45d2-9e42-73fe23da3096 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.928415] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a6524f-c007-407c-8496-b251e103b7cc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.940764] env[61824]: DEBUG nova.compute.provider_tree [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.298180] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276177, 'name': ReconfigVM_Task, 'duration_secs': 0.381562} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.298475] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 89fa4d95-b305-4490-906d-903e5823b6de/89fa4d95-b305-4490-906d-903e5823b6de.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1120.298759] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance '89fa4d95-b305-4490-906d-903e5823b6de' progress to 50 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1120.443918] env[61824]: DEBUG nova.scheduler.client.report [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1120.804828] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c0df1a-585b-4d4a-98ee-3132d3dce607 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.825199] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640b0463-741a-49d7-a33b-99287ef06bb5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.842356] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance '89fa4d95-b305-4490-906d-903e5823b6de' progress to 67 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1120.949459] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.966291] env[61824]: INFO nova.scheduler.client.report [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Deleted allocations for instance 5bf9f3b5-7084-4c5a-952b-d90c337ad10a [ 1121.417177] env[61824]: DEBUG nova.network.neutron [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Port 76ef6f05-28f7-4cff-bf0d-74bc13fdc771 binding to destination host cpu-1 is already ACTIVE {{(pid=61824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1121.474287] env[61824]: DEBUG oslo_concurrency.lockutils [None req-561ad929-e374-4dba-8c98-7371815c1fe4 tempest-AttachVolumeShelveTestJSON-1373341539 tempest-AttachVolumeShelveTestJSON-1373341539-project-member] Lock "5bf9f3b5-7084-4c5a-952b-d90c337ad10a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.302s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.440325] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "89fa4d95-b305-4490-906d-903e5823b6de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.440578] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.440778] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.387765] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1123.388030] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1123.388989] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd5e9f6-faa8-441c-9743-c7e3c032a62c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.405494] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a216457f-8ee3-41fe-b04a-627bf3a09a12 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.431173] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5/volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1123.431427] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-014acc22-88b4-455a-b479-4507bf84d26d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.449235] env[61824]: DEBUG oslo_vmware.api [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1123.449235] env[61824]: value = "task-1276180" [ 1123.449235] env[61824]: _type = "Task" [ 1123.449235] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.457611] env[61824]: DEBUG oslo_vmware.api [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276180, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.477602] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.477793] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.477979] env[61824]: DEBUG nova.network.neutron [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.959652] env[61824]: DEBUG oslo_vmware.api [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276180, 'name': ReconfigVM_Task, 'duration_secs': 0.333372} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.959939] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5/volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1123.964631] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f225085-3739-4d26-b884-0e174faf9924 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.979372] env[61824]: DEBUG oslo_vmware.api [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1123.979372] env[61824]: value = "task-1276181" [ 1123.979372] env[61824]: _type = "Task" [ 1123.979372] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.988543] env[61824]: DEBUG oslo_vmware.api [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276181, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.232527] env[61824]: DEBUG nova.network.neutron [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance_info_cache with network_info: [{"id": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "address": "fa:16:3e:d0:1e:20", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76ef6f05-28", "ovs_interfaceid": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.489311] env[61824]: DEBUG oslo_vmware.api [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276181, 'name': ReconfigVM_Task, 'duration_secs': 0.1293} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.489602] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1124.735274] env[61824]: DEBUG oslo_concurrency.lockutils [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.260144] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fe753ac-6ac5-49cf-a0b5-f4710d0ad350 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.278869] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a7652a-008f-4843-b14f-844e9e34ff2e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.285751] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance '89fa4d95-b305-4490-906d-903e5823b6de' progress to 83 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1125.524975] env[61824]: DEBUG nova.objects.instance [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'flavor' on Instance uuid 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1125.792917] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1125.793143] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f26df90-9569-4533-a3c5-ceb52bf15af9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.800830] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1125.800830] env[61824]: value = "task-1276182" [ 1125.800830] env[61824]: _type = "Task" [ 1125.800830] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.808954] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276182, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.031214] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b6e5f50b-45f9-4a7d-9883-e247d80e3df6 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.237s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.310374] env[61824]: DEBUG oslo_vmware.api [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276182, 'name': PowerOnVM_Task, 'duration_secs': 0.48475} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.310694] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1126.310868] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-2b92f321-5443-4573-96ba-8751b907f3e7 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance '89fa4d95-b305-4490-906d-903e5823b6de' progress to 100 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1126.365660] env[61824]: INFO nova.compute.manager [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Rebuilding instance [ 1126.417244] env[61824]: DEBUG nova.compute.manager [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.420026] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aee30c9-4a60-4ea5-9ae4-a5d36d55d81d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.930811] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1126.931105] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11fe6c4f-daef-4a7e-974d-6423a4cb111a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.939438] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1126.939438] env[61824]: value = "task-1276183" [ 1126.939438] env[61824]: _type = "Task" [ 1126.939438] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.948859] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.450592] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276183, 'name': PowerOffVM_Task, 'duration_secs': 0.211282} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.451039] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1127.507822] env[61824]: INFO nova.compute.manager [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Detaching volume bb2fd20c-89c6-4c6f-8752-040cc2888bf5 [ 1127.547529] env[61824]: INFO nova.virt.block_device [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Attempting to driver detach volume bb2fd20c-89c6-4c6f-8752-040cc2888bf5 from mountpoint /dev/sdb [ 1127.547777] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1127.547969] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1127.549077] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0beb9ab-da18-4561-a0ea-3e8c9e1e5319 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.572655] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff896b0-07ce-4ba0-af43-ba9bd65b5366 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.579589] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6778e2d2-8278-4d29-b697-73cadcaeeaca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.599887] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8358f84e-2139-410a-938f-62a136b714e5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.614057] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] The volume has not been displaced from its original location: [datastore1] volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5/volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1127.619281] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1127.619559] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3dd7d8a2-0723-424c-8961-311cd0eca225 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.636999] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1127.636999] env[61824]: value = "task-1276184" [ 1127.636999] env[61824]: _type = "Task" [ 1127.636999] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.644747] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276184, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.146678] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276184, 'name': ReconfigVM_Task, 'duration_secs': 0.190278} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.147093] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1128.151662] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1cf5143-dc57-40d6-b5d2-31f6664560c9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.166686] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1128.166686] env[61824]: value = "task-1276185" [ 1128.166686] env[61824]: _type = "Task" [ 1128.166686] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.174674] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.199665] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "89fa4d95-b305-4490-906d-903e5823b6de" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.199918] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.200148] env[61824]: DEBUG nova.compute.manager [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Going to confirm migration 3 {{(pid=61824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1128.678460] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276185, 'name': ReconfigVM_Task, 'duration_secs': 0.156352} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.678760] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1128.735048] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.735259] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquired lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.735436] env[61824]: DEBUG nova.network.neutron [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1128.735623] env[61824]: DEBUG nova.objects.instance [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'info_cache' on Instance uuid 89fa4d95-b305-4490-906d-903e5823b6de {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.730634] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1129.731058] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89468389-23d0-4773-8f5b-0e94c1f97e7d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.738570] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1129.738570] env[61824]: value = "task-1276186" [ 1129.738570] env[61824]: _type = "Task" [ 1129.738570] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.748719] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] VM already powered off {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1129.748927] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1129.749150] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1129.749876] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93eb3ee5-a000-4bc9-ae43-1f427bb100c0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.767260] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e098ca75-e3e1-4613-a98d-abc40703a3da {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.774782] env[61824]: WARNING nova.virt.vmwareapi.driver [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1129.775076] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1129.775792] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de734844-237e-43f3-889e-39a19d866ca6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.781753] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1129.781974] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ec9dfaf-f4a6-480f-b46e-4c511d8dd0b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.843826] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.844078] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.844270] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleting the datastore file [datastore2] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.844546] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f38e6d37-1282-4a2a-b780-9f1d94b2a517 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.851154] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1129.851154] env[61824]: value = "task-1276188" [ 1129.851154] env[61824]: _type = "Task" [ 1129.851154] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.861466] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276188, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.944143] env[61824]: DEBUG nova.network.neutron [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance_info_cache with network_info: [{"id": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "address": "fa:16:3e:d0:1e:20", "network": {"id": "b663c01b-4bbb-4fa7-b4c1-046c2ad69aef", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-352053583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "629276280a234e7a8b810e6e74779fff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76ef6f05-28", "ovs_interfaceid": "76ef6f05-28f7-4cff-bf0d-74bc13fdc771", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.363411] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276188, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161448} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.364700] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1130.364700] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1130.364700] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1130.446746] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Releasing lock "refresh_cache-89fa4d95-b305-4490-906d-903e5823b6de" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.447085] env[61824]: DEBUG nova.objects.instance [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lazy-loading 'migration_context' on Instance uuid 89fa4d95-b305-4490-906d-903e5823b6de {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1130.868991] env[61824]: INFO nova.virt.block_device [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Booting with volume bb2fd20c-89c6-4c6f-8752-040cc2888bf5 at /dev/sdb [ 1130.901942] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e50cb74-d270-4e76-9138-b7f7e755c646 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.910207] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5f3fdc-149c-4b09-9e09-c7607a766a1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.935665] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-12435e6b-a936-43d5-a188-5480cc08b6d0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.943018] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912075a5-0032-44a4-b877-71d34ef94704 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.953374] env[61824]: DEBUG nova.objects.base [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Object Instance<89fa4d95-b305-4490-906d-903e5823b6de> lazy-loaded attributes: info_cache,migration_context {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1130.954113] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45eb2a7a-39ad-4b4e-8635-78a404cd72b7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.976928] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a04f590c-b2f6-4bb6-8439-2b87d4e0a42a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.979413] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb33caf-694d-4519-86ae-92d96f06f224 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.984063] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1130.984063] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52a63912-6659-997d-98ed-320f9a4a08c3" [ 1130.984063] env[61824]: _type = "Task" [ 1130.984063] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.988607] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026b788c-8f23-456d-8c05-306bb488e6d6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.998691] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52a63912-6659-997d-98ed-320f9a4a08c3, 'name': SearchDatastore_Task, 'duration_secs': 0.006144} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.002596] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.002777] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.004100] env[61824]: DEBUG nova.virt.block_device [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updating existing volume attachment record: d1d3c6ec-d905-42cf-8e71-53aa22112894 {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1131.596166] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c0147a-b07d-4c03-8e3b-d86b2862cc4a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.603980] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012b17e3-25c7-44f5-a736-0c1e63743611 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.634018] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7f970b-d0d0-499a-b574-e5b729675769 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.640789] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312f177a-e54c-4fe2-bad2-235654ff6256 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.653643] env[61824]: DEBUG nova.compute.provider_tree [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.157515] env[61824]: DEBUG nova.scheduler.client.report [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1132.646075] env[61824]: DEBUG oslo_concurrency.lockutils [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "b8db876d-e430-4c5f-a439-0601dbf4598d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.646400] env[61824]: DEBUG oslo_concurrency.lockutils [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.104593] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1133.104851] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1133.105027] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1133.105219] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1133.105370] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1133.105521] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1133.105734] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1133.105902] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1133.106092] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1133.106265] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1133.106446] env[61824]: DEBUG nova.virt.hardware [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1133.107338] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce5a754-1e94-4b3b-a8f0-7c279f59802d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.114786] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e006a7cf-835f-471f-9391-54615f360e12 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.129083] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:ae:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7afd60b2-5996-4b14-9a2a-2c271dc1394c', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1133.136380] env[61824]: DEBUG oslo.service.loopingcall [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.136611] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1133.136808] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3009b5d4-03bd-4eb9-b30c-1c371c9711e2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.150889] env[61824]: INFO nova.compute.manager [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Detaching volume ccf8167d-4a22-469a-9a8b-c7baa195387f [ 1133.158897] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1133.158897] env[61824]: value = "task-1276189" [ 1133.158897] env[61824]: _type = "Task" [ 1133.158897] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.166895] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.164s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.169526] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276189, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.182317] env[61824]: INFO nova.virt.block_device [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Attempting to driver detach volume ccf8167d-4a22-469a-9a8b-c7baa195387f from mountpoint /dev/sdb [ 1133.182721] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1133.183089] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274245', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'name': 'volume-ccf8167d-4a22-469a-9a8b-c7baa195387f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8db876d-e430-4c5f-a439-0601dbf4598d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'serial': 'ccf8167d-4a22-469a-9a8b-c7baa195387f'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1133.183994] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef65e0c-bcc3-45fe-a858-33807f04da3c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.207019] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83fe7d5-b730-4616-bc42-c56d98219619 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.215011] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced4cb44-033c-412e-930d-e1dd5271c751 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.235286] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b332a9-ad85-4b9a-8b63-67e1fc9c75ce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.249800] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] The volume has not been displaced from its original location: [datastore1] volume-ccf8167d-4a22-469a-9a8b-c7baa195387f/volume-ccf8167d-4a22-469a-9a8b-c7baa195387f.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1133.255078] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.255677] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e2ab7d5-ba54-4c64-b67e-aa986d35e5ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.272951] env[61824]: DEBUG oslo_vmware.api [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1133.272951] env[61824]: value = "task-1276190" [ 1133.272951] env[61824]: _type = "Task" [ 1133.272951] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.280812] env[61824]: DEBUG oslo_vmware.api [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276190, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.673319] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276189, 'name': CreateVM_Task, 'duration_secs': 0.28595} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.675388] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1133.676233] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.676583] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.676761] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1133.676951] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1797adcc-9164-4aec-a893-2218dc1fdfce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.681455] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1133.681455] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52ba12d8-7d7c-2bc7-5f4e-2fa1172496ba" [ 1133.681455] env[61824]: _type = "Task" [ 1133.681455] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.690142] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ba12d8-7d7c-2bc7-5f4e-2fa1172496ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.721934] env[61824]: INFO nova.scheduler.client.report [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted allocation for migration 559c6e8d-7c7e-4560-a123-167e6b42318f [ 1133.782529] env[61824]: DEBUG oslo_vmware.api [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276190, 'name': ReconfigVM_Task, 'duration_secs': 0.238768} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.782828] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1133.787522] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4834fba-572a-426f-9530-c2da059aaa60 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.801946] env[61824]: DEBUG oslo_vmware.api [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1133.801946] env[61824]: value = "task-1276191" [ 1133.801946] env[61824]: _type = "Task" [ 1133.801946] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.809916] env[61824]: DEBUG oslo_vmware.api [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.191627] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52ba12d8-7d7c-2bc7-5f4e-2fa1172496ba, 'name': SearchDatastore_Task, 'duration_secs': 0.00902} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.191980] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1134.192173] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1134.192424] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1134.192575] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.192781] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1134.193065] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-555a6bd4-a8aa-45f1-8e86-7f875a5eba7c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.201112] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1134.201267] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1134.201983] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d16dec4f-9f73-42d5-b55d-34416297be7e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.207076] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1134.207076] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]524ab710-2c26-3fdd-7ef8-f359afff4556" [ 1134.207076] env[61824]: _type = "Task" [ 1134.207076] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.214105] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524ab710-2c26-3fdd-7ef8-f359afff4556, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.226500] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.027s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.312070] env[61824]: DEBUG oslo_vmware.api [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276191, 'name': ReconfigVM_Task, 'duration_secs': 0.131401} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.312382] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274245', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'name': 'volume-ccf8167d-4a22-469a-9a8b-c7baa195387f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b8db876d-e430-4c5f-a439-0601dbf4598d', 'attached_at': '', 'detached_at': '', 'volume_id': 'ccf8167d-4a22-469a-9a8b-c7baa195387f', 'serial': 'ccf8167d-4a22-469a-9a8b-c7baa195387f'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1134.717727] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]524ab710-2c26-3fdd-7ef8-f359afff4556, 'name': SearchDatastore_Task, 'duration_secs': 0.008109} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.718549] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-788789dd-df78-4fc0-85ed-944826588d30 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.723821] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1134.723821] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]522f195e-8c01-e241-1202-b8c0913efa3f" [ 1134.723821] env[61824]: _type = "Task" [ 1134.723821] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.731625] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522f195e-8c01-e241-1202-b8c0913efa3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.851398] env[61824]: DEBUG nova.objects.instance [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lazy-loading 'flavor' on Instance uuid b8db876d-e430-4c5f-a439-0601dbf4598d {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.234193] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]522f195e-8c01-e241-1202-b8c0913efa3f, 'name': SearchDatastore_Task, 'duration_secs': 0.008727} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.234494] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.234726] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1135.235015] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5159ae1-954a-42ce-b129-a4b6a7d7a5a0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.241534] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1135.241534] env[61824]: value = "task-1276192" [ 1135.241534] env[61824]: _type = "Task" [ 1135.241534] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.249242] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276192, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.720251] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "89fa4d95-b305-4490-906d-903e5823b6de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.720644] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.720868] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "89fa4d95-b305-4490-906d-903e5823b6de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.721085] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.721266] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.723454] env[61824]: INFO nova.compute.manager [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Terminating instance [ 1135.725199] env[61824]: DEBUG nova.compute.manager [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1135.725397] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1135.726278] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcca3bc3-b5af-420d-91d4-e12bc844a468 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.733698] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1135.733941] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6748c23-0244-4b54-aa2a-478ff5638a9b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.740142] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1135.740142] env[61824]: value = "task-1276193" [ 1135.740142] env[61824]: _type = "Task" [ 1135.740142] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.750136] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276193, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.753113] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276192, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.426601} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.753344] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore1] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1135.753549] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1135.753773] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7df2118b-5c27-40d0-8780-a15866b246d9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.759820] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1135.759820] env[61824]: value = "task-1276194" [ 1135.759820] env[61824]: _type = "Task" [ 1135.759820] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.767160] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276194, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.858974] env[61824]: DEBUG oslo_concurrency.lockutils [None req-41643891-1c3d-4cb5-ac5b-19b77776015f tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.212s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.250367] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276193, 'name': PowerOffVM_Task, 'duration_secs': 0.216804} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.250709] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1136.250865] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1136.251127] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6087f697-e071-4039-94ad-cda81a5f9d24 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.268069] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276194, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088557} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.268313] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1136.269021] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79fcb734-7078-4ca8-914c-ba1287bd4986 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.290737] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.291081] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab84d8fd-12e4-42db-b6dc-1773bc1b2384 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.314042] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1136.314042] env[61824]: value = "task-1276196" [ 1136.314042] env[61824]: _type = "Task" [ 1136.314042] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.314042] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1136.314042] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1136.314325] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleting the datastore file [datastore1] 89fa4d95-b305-4490-906d-903e5823b6de {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1136.314468] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3fb3964-b37f-4ec7-8e3a-f7c7b97c3900 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.324306] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276196, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.325502] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for the task: (returnval){ [ 1136.325502] env[61824]: value = "task-1276197" [ 1136.325502] env[61824]: _type = "Task" [ 1136.325502] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.332651] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276197, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.825587] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276196, 'name': ReconfigVM_Task, 'duration_secs': 0.290739} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.825587] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72/4bdef3c0-7643-48d8-a067-cdbb7e2f0c72.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1136.826645] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'disk_bus': None, 'boot_index': 0, 'guest_format': None, 'encryption_secret_uuid': None, 'encryption_options': None, 'size': 0, 'encrypted': False, 'device_name': '/dev/sda', 'encryption_format': None, 'device_type': 'disk', 'image_id': '11b2670a-52f3-4a0e-8363-54ca7a7f5264'}], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'boot_index': None, 'guest_format': None, 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'}, 'mount_device': '/dev/sdb', 'attachment_id': 'd1d3c6ec-d905-42cf-8e71-53aa22112894', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1136.826845] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1136.827048] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1136.827874] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259cc7db-5ba7-4a40-bc1f-f3288158f430 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.837772] env[61824]: DEBUG oslo_vmware.api [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Task: {'id': task-1276197, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.221006} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.847928] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.848133] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1136.848315] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1136.848492] env[61824]: INFO nova.compute.manager [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1136.848735] env[61824]: DEBUG oslo.service.loopingcall [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.849087] env[61824]: DEBUG nova.compute.manager [-] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1136.849189] env[61824]: DEBUG nova.network.neutron [-] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1136.851193] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33680b4d-e56f-44c8-8155-7120860ac39d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.875373] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5/volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.875644] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc9aa93f-3806-43d3-9af4-5179838d3224 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.895902] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1136.895902] env[61824]: value = "task-1276198" [ 1136.895902] env[61824]: _type = "Task" [ 1136.895902] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.903661] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276198, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.966443] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "b8db876d-e430-4c5f-a439-0601dbf4598d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.966731] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.966947] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "b8db876d-e430-4c5f-a439-0601dbf4598d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.967191] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.967375] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.969584] env[61824]: INFO nova.compute.manager [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Terminating instance [ 1136.971488] env[61824]: DEBUG nova.compute.manager [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1136.971688] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1136.972581] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc57e0ef-e471-4ac3-bf20-8ca2da6ad600 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.981579] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1136.981930] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e809f91f-93c3-46c9-b1cf-b4f33b7ab6b2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.989494] env[61824]: DEBUG oslo_vmware.api [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1136.989494] env[61824]: value = "task-1276199" [ 1136.989494] env[61824]: _type = "Task" [ 1136.989494] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.001268] env[61824]: DEBUG oslo_vmware.api [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.204473] env[61824]: DEBUG nova.compute.manager [req-28541ec2-422b-4547-be8f-3beca32b467c req-6cea2443-5808-457b-8bb1-33e211ee707b service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Received event network-vif-deleted-76ef6f05-28f7-4cff-bf0d-74bc13fdc771 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1137.204473] env[61824]: INFO nova.compute.manager [req-28541ec2-422b-4547-be8f-3beca32b467c req-6cea2443-5808-457b-8bb1-33e211ee707b service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Neutron deleted interface 76ef6f05-28f7-4cff-bf0d-74bc13fdc771; detaching it from the instance and deleting it from the info cache [ 1137.204473] env[61824]: DEBUG nova.network.neutron [req-28541ec2-422b-4547-be8f-3beca32b467c req-6cea2443-5808-457b-8bb1-33e211ee707b service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.406137] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276198, 'name': ReconfigVM_Task, 'duration_secs': 0.324267} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.406475] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5/volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1137.411208] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2124ec14-e04a-4a96-a3f9-ea4eb67d7e41 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.426228] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1137.426228] env[61824]: value = "task-1276200" [ 1137.426228] env[61824]: _type = "Task" [ 1137.426228] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.433738] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276200, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.499449] env[61824]: DEBUG oslo_vmware.api [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276199, 'name': PowerOffVM_Task, 'duration_secs': 0.238256} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.499720] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1137.499892] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1137.500153] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-550ddca3-75db-411b-af41-fcaaf24a281e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.556607] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1137.556898] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1137.557070] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleting the datastore file [datastore2] b8db876d-e430-4c5f-a439-0601dbf4598d {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1137.557354] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39e6b066-ad45-41ed-bc99-b5b3fcdc9013 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.563035] env[61824]: DEBUG oslo_vmware.api [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1137.563035] env[61824]: value = "task-1276202" [ 1137.563035] env[61824]: _type = "Task" [ 1137.563035] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.570358] env[61824]: DEBUG oslo_vmware.api [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.584763] env[61824]: DEBUG nova.network.neutron [-] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.707771] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1eef3874-c482-40a5-bd64-c28bbe674e3a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.717177] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba9c355-cdb7-4d0c-bf15-42fa7e1c5ac0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.741169] env[61824]: DEBUG nova.compute.manager [req-28541ec2-422b-4547-be8f-3beca32b467c req-6cea2443-5808-457b-8bb1-33e211ee707b service nova] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Detach interface failed, port_id=76ef6f05-28f7-4cff-bf0d-74bc13fdc771, reason: Instance 89fa4d95-b305-4490-906d-903e5823b6de could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1137.936335] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276200, 'name': ReconfigVM_Task, 'duration_secs': 0.13591} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.936642] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1137.937247] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e8a7518-58a7-4688-aee0-7b7fae7b5193 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.943971] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1137.943971] env[61824]: value = "task-1276203" [ 1137.943971] env[61824]: _type = "Task" [ 1137.943971] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.952141] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276203, 'name': Rename_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.072073] env[61824]: DEBUG oslo_vmware.api [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154346} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.072270] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.072441] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1138.072618] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1138.072793] env[61824]: INFO nova.compute.manager [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1138.073132] env[61824]: DEBUG oslo.service.loopingcall [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.073308] env[61824]: DEBUG nova.compute.manager [-] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1138.073402] env[61824]: DEBUG nova.network.neutron [-] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1138.087651] env[61824]: INFO nova.compute.manager [-] [instance: 89fa4d95-b305-4490-906d-903e5823b6de] Took 1.24 seconds to deallocate network for instance. [ 1138.453957] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276203, 'name': Rename_Task, 'duration_secs': 0.144477} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.454380] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1138.454444] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0fa2618-88ce-4598-8f47-8957b7409cdf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.460171] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1138.460171] env[61824]: value = "task-1276204" [ 1138.460171] env[61824]: _type = "Task" [ 1138.460171] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.467316] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276204, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.595229] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.599021] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.599021] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.620343] env[61824]: INFO nova.scheduler.client.report [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Deleted allocations for instance 89fa4d95-b305-4490-906d-903e5823b6de [ 1138.970235] env[61824]: DEBUG oslo_vmware.api [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276204, 'name': PowerOnVM_Task, 'duration_secs': 0.492415} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.970456] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1138.970654] env[61824]: DEBUG nova.compute.manager [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1138.971453] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a847382-d479-453d-b984-2574689a3541 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.028834] env[61824]: DEBUG nova.network.neutron [-] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.129423] env[61824]: DEBUG oslo_concurrency.lockutils [None req-097d0678-422f-46d3-b56b-c57eae377f17 tempest-DeleteServersTestJSON-1224720787 tempest-DeleteServersTestJSON-1224720787-project-member] Lock "89fa4d95-b305-4490-906d-903e5823b6de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.409s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.229710] env[61824]: DEBUG nova.compute.manager [req-fd1f4a52-172d-4cea-a680-16e5182875fd req-81024da0-1fe0-4dac-895d-4ac5754d52e0 service nova] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Received event network-vif-deleted-c75f0bf1-26d7-46fd-a861-f3ddbf753d2d {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1139.490028] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.490396] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.490673] env[61824]: DEBUG nova.objects.instance [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61824) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1139.531525] env[61824]: INFO nova.compute.manager [-] [instance: b8db876d-e430-4c5f-a439-0601dbf4598d] Took 1.46 seconds to deallocate network for instance. [ 1140.037956] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.500408] env[61824]: DEBUG oslo_concurrency.lockutils [None req-3902386d-4937-40fb-a414-3f68597f344b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.501553] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.464s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.501776] env[61824]: DEBUG nova.objects.instance [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lazy-loading 'resources' on Instance uuid b8db876d-e430-4c5f-a439-0601dbf4598d {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.568106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.568360] env[61824]: DEBUG oslo_concurrency.lockutils [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.070403] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf8b1e2-ffe9-4ac3-8ab9-f78f709d4d61 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.073536] env[61824]: INFO nova.compute.manager [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Detaching volume bb2fd20c-89c6-4c6f-8752-040cc2888bf5 [ 1141.082016] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cd68a5c-b387-4be1-99a3-eda279ec841e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.113754] env[61824]: INFO nova.virt.block_device [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Attempting to driver detach volume bb2fd20c-89c6-4c6f-8752-040cc2888bf5 from mountpoint /dev/sdb [ 1141.113987] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1141.114191] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1141.114940] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb06c306-813f-4de4-9d18-fd0617f606ca {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.117929] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e27621-1f0d-4016-82d7-210cc7858e37 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.139858] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856239d8-1f02-48d3-a6ec-7e7b9ebcd09f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.143853] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73e3e2ad-c399-4b2b-a9a5-e8862b9cfad1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.156526] env[61824]: DEBUG nova.compute.provider_tree [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.159607] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937a180d-c583-4a8c-b29f-bc01966f029a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.180014] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52e81c2-6366-48dd-9389-40287ea0e39f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.194409] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] The volume has not been displaced from its original location: [datastore1] volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5/volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1141.199519] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1141.199798] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-313d9016-e48c-4177-85e2-86e94e893cfa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.216615] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1141.216615] env[61824]: value = "task-1276206" [ 1141.216615] env[61824]: _type = "Task" [ 1141.216615] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.224823] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276206, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.660544] env[61824]: DEBUG nova.scheduler.client.report [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1141.725580] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276206, 'name': ReconfigVM_Task, 'duration_secs': 0.200275} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.725870] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1141.731270] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-79551428-be37-4c08-b14f-a1fa29fd4f87 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.746050] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1141.746050] env[61824]: value = "task-1276207" [ 1141.746050] env[61824]: _type = "Task" [ 1141.746050] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.753307] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276207, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.166239] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.183728] env[61824]: INFO nova.scheduler.client.report [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleted allocations for instance b8db876d-e430-4c5f-a439-0601dbf4598d [ 1142.255586] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276207, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.691680] env[61824]: DEBUG oslo_concurrency.lockutils [None req-60f3c58c-89ad-4c2b-8ea4-02fbc529c389 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "b8db876d-e430-4c5f-a439-0601dbf4598d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.725s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.756763] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276207, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.256704] env[61824]: DEBUG oslo_vmware.api [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276207, 'name': ReconfigVM_Task, 'duration_secs': 1.137072} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.257049] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274249', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'name': 'volume-bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4bdef3c0-7643-48d8-a067-cdbb7e2f0c72', 'attached_at': '', 'detached_at': '', 'volume_id': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5', 'serial': 'bb2fd20c-89c6-4c6f-8752-040cc2888bf5'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1143.801296] env[61824]: DEBUG nova.objects.instance [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'flavor' on Instance uuid 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1144.809704] env[61824]: DEBUG oslo_concurrency.lockutils [None req-383ba4e8-e780-4bc8-baf5-367c1432f314 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.241s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.847819] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.848220] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.848495] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.848699] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.848976] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.851193] env[61824]: INFO nova.compute.manager [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Terminating instance [ 1145.853021] env[61824]: DEBUG nova.compute.manager [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1145.853292] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1145.854169] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4170a5a0-0f70-44de-aefa-d2860b9e7136 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.861968] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1145.862252] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38599ee4-c3a1-4beb-ac35-ccdee023263b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.868398] env[61824]: DEBUG oslo_vmware.api [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1145.868398] env[61824]: value = "task-1276209" [ 1145.868398] env[61824]: _type = "Task" [ 1145.868398] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.877258] env[61824]: DEBUG oslo_vmware.api [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.378932] env[61824]: DEBUG oslo_vmware.api [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276209, 'name': PowerOffVM_Task, 'duration_secs': 0.19456} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.379237] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1146.379413] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1146.379654] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ce49d1b-d643-402f-bd97-9d6b598d1dbf {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.435027] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1146.435027] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1146.435173] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleting the datastore file [datastore1] 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1146.435417] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cbbe0bbf-76a8-46be-9a50-de2bd685b4bb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.441475] env[61824]: DEBUG oslo_vmware.api [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1146.441475] env[61824]: value = "task-1276211" [ 1146.441475] env[61824]: _type = "Task" [ 1146.441475] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.449014] env[61824]: DEBUG oslo_vmware.api [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276211, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.951613] env[61824]: DEBUG oslo_vmware.api [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276211, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125285} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.951935] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1146.952171] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1146.952358] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1146.952589] env[61824]: INFO nova.compute.manager [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1146.952866] env[61824]: DEBUG oslo.service.loopingcall [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1146.953103] env[61824]: DEBUG nova.compute.manager [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1146.953249] env[61824]: DEBUG nova.network.neutron [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1147.347086] env[61824]: DEBUG nova.compute.manager [req-dcf6dbeb-d457-4999-ac89-c13e77f68e9c req-e7e88831-e065-4cbf-8195-c4084ff9a31c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Received event network-vif-deleted-7afd60b2-5996-4b14-9a2a-2c271dc1394c {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.347320] env[61824]: INFO nova.compute.manager [req-dcf6dbeb-d457-4999-ac89-c13e77f68e9c req-e7e88831-e065-4cbf-8195-c4084ff9a31c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Neutron deleted interface 7afd60b2-5996-4b14-9a2a-2c271dc1394c; detaching it from the instance and deleting it from the info cache [ 1147.347498] env[61824]: DEBUG nova.network.neutron [req-dcf6dbeb-d457-4999-ac89-c13e77f68e9c req-e7e88831-e065-4cbf-8195-c4084ff9a31c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.827588] env[61824]: DEBUG nova.network.neutron [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.850248] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-503f40fe-a5ff-414e-b83c-b2313d1ef31c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.860139] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a85609-01bb-4c46-828f-14b92689f1df {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.883319] env[61824]: DEBUG nova.compute.manager [req-dcf6dbeb-d457-4999-ac89-c13e77f68e9c req-e7e88831-e065-4cbf-8195-c4084ff9a31c service nova] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Detach interface failed, port_id=7afd60b2-5996-4b14-9a2a-2c271dc1394c, reason: Instance 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1148.032262] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.032645] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.032734] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.032922] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.033112] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.035262] env[61824]: INFO nova.compute.manager [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Terminating instance [ 1148.036972] env[61824]: DEBUG nova.compute.manager [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1148.037183] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1148.037996] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8706546-2ff7-4b35-883f-057e2bd18b33 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.045373] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1148.045597] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5dcbbdd-5be1-4df2-b4a3-46b07cd02a1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.051776] env[61824]: DEBUG oslo_vmware.api [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1148.051776] env[61824]: value = "task-1276212" [ 1148.051776] env[61824]: _type = "Task" [ 1148.051776] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.060432] env[61824]: DEBUG oslo_vmware.api [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.331101] env[61824]: INFO nova.compute.manager [-] [instance: 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72] Took 1.38 seconds to deallocate network for instance. [ 1148.564036] env[61824]: DEBUG oslo_vmware.api [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276212, 'name': PowerOffVM_Task, 'duration_secs': 0.186414} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.564350] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1148.564475] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1148.564738] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dfc2d40-020b-49b8-be24-0d9ede6c679d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.620532] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1148.620764] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1148.620881] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleting the datastore file [datastore2] 1ddb6d7a-d8e1-492d-84c9-d61e27faa105 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1148.621193] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-752d718f-cb10-4581-9d8f-536c545d8525 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.626608] env[61824]: DEBUG oslo_vmware.api [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1148.626608] env[61824]: value = "task-1276214" [ 1148.626608] env[61824]: _type = "Task" [ 1148.626608] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.634298] env[61824]: DEBUG oslo_vmware.api [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276214, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.837021] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.837339] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.837580] env[61824]: DEBUG nova.objects.instance [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'resources' on Instance uuid 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1149.136371] env[61824]: DEBUG oslo_vmware.api [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276214, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164972} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.136745] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1149.136825] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1149.137035] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1149.137226] env[61824]: INFO nova.compute.manager [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1149.137469] env[61824]: DEBUG oslo.service.loopingcall [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.137666] env[61824]: DEBUG nova.compute.manager [-] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1149.137759] env[61824]: DEBUG nova.network.neutron [-] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1149.373439] env[61824]: DEBUG nova.compute.manager [req-e96bb3e2-163b-4000-8117-3b1cc6d973fe req-019193b2-537e-4bd8-b908-a5e5b2bce240 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Received event network-vif-deleted-3b69bd09-a533-499d-b63a-e125b6c295fd {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1149.373674] env[61824]: INFO nova.compute.manager [req-e96bb3e2-163b-4000-8117-3b1cc6d973fe req-019193b2-537e-4bd8-b908-a5e5b2bce240 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Neutron deleted interface 3b69bd09-a533-499d-b63a-e125b6c295fd; detaching it from the instance and deleting it from the info cache [ 1149.373828] env[61824]: DEBUG nova.network.neutron [req-e96bb3e2-163b-4000-8117-3b1cc6d973fe req-019193b2-537e-4bd8-b908-a5e5b2bce240 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.407950] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3dfe529-555a-4f7a-8802-c8c6d93d5420 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.415452] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1896d6-db2c-49d9-a665-8741df507112 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.445045] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea7c7d4-15e3-4f29-abca-7b6b61b0f503 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.451449] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8541b3-3928-402b-8a2d-918553a64c8d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.465433] env[61824]: DEBUG nova.compute.provider_tree [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.850507] env[61824]: DEBUG nova.network.neutron [-] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.878699] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-659d4bd5-a347-437d-a12b-50c138a1c561 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.889305] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03b52ea-2e3f-48bb-b880-fb1d506cb23f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.912957] env[61824]: DEBUG nova.compute.manager [req-e96bb3e2-163b-4000-8117-3b1cc6d973fe req-019193b2-537e-4bd8-b908-a5e5b2bce240 service nova] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Detach interface failed, port_id=3b69bd09-a533-499d-b63a-e125b6c295fd, reason: Instance 1ddb6d7a-d8e1-492d-84c9-d61e27faa105 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1149.967901] env[61824]: DEBUG nova.scheduler.client.report [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1150.352928] env[61824]: INFO nova.compute.manager [-] [instance: 1ddb6d7a-d8e1-492d-84c9-d61e27faa105] Took 1.21 seconds to deallocate network for instance. [ 1150.473131] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.636s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.495087] env[61824]: INFO nova.scheduler.client.report [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted allocations for instance 4bdef3c0-7643-48d8-a067-cdbb7e2f0c72 [ 1150.858589] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.858874] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.859129] env[61824]: DEBUG nova.objects.instance [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lazy-loading 'resources' on Instance uuid 1ddb6d7a-d8e1-492d-84c9-d61e27faa105 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.002340] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1e834ead-1e90-4b99-8685-e69fa2d33fbd tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "4bdef3c0-7643-48d8-a067-cdbb7e2f0c72" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.154s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.411054] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0a8a3f-2fa0-481b-9b38-469d9818763d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.418296] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6496607f-8b58-4ec8-aed9-2aceaa02bc24 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.449243] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d37efaf-7817-405d-b48a-801dc57f2944 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.455662] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00bb99f-bba8-42f7-882f-c8540a078e1d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.468188] env[61824]: DEBUG nova.compute.provider_tree [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.971434] env[61824]: DEBUG nova.scheduler.client.report [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1152.476257] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.617s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.478206] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.478427] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.495292] env[61824]: INFO nova.scheduler.client.report [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleted allocations for instance 1ddb6d7a-d8e1-492d-84c9-d61e27faa105 [ 1152.981170] env[61824]: DEBUG nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1153.002703] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1fd16aec-6555-4a25-8a40-54bbf1ed325e tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "1ddb6d7a-d8e1-492d-84c9-d61e27faa105" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.970s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.503014] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.503392] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.504760] env[61824]: INFO nova.compute.claims [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1153.562553] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.562809] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.563032] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.563228] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.563466] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.565482] env[61824]: INFO nova.compute.manager [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Terminating instance [ 1153.567283] env[61824]: DEBUG nova.compute.manager [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1153.567476] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1153.568328] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae331a97-b132-4689-a746-c1a6ac1775bd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.576422] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1153.576659] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a284205-20b2-4dd5-8e51-90c2de085f38 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.583772] env[61824]: DEBUG oslo_vmware.api [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1153.583772] env[61824]: value = "task-1276215" [ 1153.583772] env[61824]: _type = "Task" [ 1153.583772] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.591394] env[61824]: DEBUG oslo_vmware.api [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276215, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.092843] env[61824]: DEBUG oslo_vmware.api [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276215, 'name': PowerOffVM_Task, 'duration_secs': 0.175683} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.093158] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1154.093351] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1154.093592] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1c665112-eeba-4531-8162-ceef1f477cbb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.154648] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1154.154870] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Deleting contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1154.155068] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleting the datastore file [datastore1] 4700de90-7230-43a5-87ae-f2c7ed9d4ea6 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.155337] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fbe3c2d2-6428-4445-ab89-6509ace2716e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.161184] env[61824]: DEBUG oslo_vmware.api [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for the task: (returnval){ [ 1154.161184] env[61824]: value = "task-1276217" [ 1154.161184] env[61824]: _type = "Task" [ 1154.161184] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.168206] env[61824]: DEBUG oslo_vmware.api [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276217, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.560098] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962ff894-2c11-4aa1-8d3a-a4ca36e89950 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.567091] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48b0920-99e2-4645-823e-3850275b696c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.595550] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc0452d-4e15-49af-8ecd-f2ef678a77c3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.601956] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378920ce-c90a-4195-ada4-4056690e1e70 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.615734] env[61824]: DEBUG nova.compute.provider_tree [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.670573] env[61824]: DEBUG oslo_vmware.api [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Task: {'id': task-1276217, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134519} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.670834] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1154.671054] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Deleted contents of the VM from datastore datastore1 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1154.671276] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1154.671458] env[61824]: INFO nova.compute.manager [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1154.671698] env[61824]: DEBUG oslo.service.loopingcall [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1154.671889] env[61824]: DEBUG nova.compute.manager [-] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1154.671986] env[61824]: DEBUG nova.network.neutron [-] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1154.902243] env[61824]: DEBUG nova.compute.manager [req-e0872bc3-dc33-43ad-81f3-624e7b0b9e69 req-42c017c2-5fc0-47b2-add2-db7052d496e1 service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Received event network-vif-deleted-2a20cdd6-6c0a-46d2-b5a7-1b78e2476561 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1154.902457] env[61824]: INFO nova.compute.manager [req-e0872bc3-dc33-43ad-81f3-624e7b0b9e69 req-42c017c2-5fc0-47b2-add2-db7052d496e1 service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Neutron deleted interface 2a20cdd6-6c0a-46d2-b5a7-1b78e2476561; detaching it from the instance and deleting it from the info cache [ 1154.902632] env[61824]: DEBUG nova.network.neutron [req-e0872bc3-dc33-43ad-81f3-624e7b0b9e69 req-42c017c2-5fc0-47b2-add2-db7052d496e1 service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.119066] env[61824]: DEBUG nova.scheduler.client.report [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.379306] env[61824]: DEBUG nova.network.neutron [-] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.405109] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f0003a7b-8bc5-4a6f-a5c1-cb3f5154a945 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.414826] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8336850f-43a0-4b36-89d2-653a59da729b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.437578] env[61824]: DEBUG nova.compute.manager [req-e0872bc3-dc33-43ad-81f3-624e7b0b9e69 req-42c017c2-5fc0-47b2-add2-db7052d496e1 service nova] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Detach interface failed, port_id=2a20cdd6-6c0a-46d2-b5a7-1b78e2476561, reason: Instance 4700de90-7230-43a5-87ae-f2c7ed9d4ea6 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1155.624649] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.121s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.625235] env[61824]: DEBUG nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1155.881637] env[61824]: INFO nova.compute.manager [-] [instance: 4700de90-7230-43a5-87ae-f2c7ed9d4ea6] Took 1.21 seconds to deallocate network for instance. [ 1156.130473] env[61824]: DEBUG nova.compute.utils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1156.132337] env[61824]: DEBUG nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1156.132553] env[61824]: DEBUG nova.network.neutron [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1156.172405] env[61824]: DEBUG nova.policy [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad4e8224518048e583d27b94721b3c22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd21256931aa54010b72beca4fb798f63', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1156.387992] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.388303] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.388532] env[61824]: DEBUG nova.objects.instance [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lazy-loading 'resources' on Instance uuid 4700de90-7230-43a5-87ae-f2c7ed9d4ea6 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.409026] env[61824]: DEBUG nova.network.neutron [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Successfully created port: 1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1156.635713] env[61824]: DEBUG nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1156.946781] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da46d4ad-a202-4162-b27c-b8235ef1ff28 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.954620] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13eec684-552c-4dca-9391-68d7bdd97cb1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.984740] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614cfcc1-a07d-4f53-a26b-2bb0a0bbb38d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.991781] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb23dd7a-a007-476b-9565-7efd1a9bf4f5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.004613] env[61824]: DEBUG nova.compute.provider_tree [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.507914] env[61824]: DEBUG nova.scheduler.client.report [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1157.645047] env[61824]: DEBUG nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1157.670779] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1157.671073] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1157.671265] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1157.671460] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1157.671611] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1157.671761] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1157.671969] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1157.672193] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1157.672385] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1157.672560] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1157.672735] env[61824]: DEBUG nova.virt.hardware [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1157.673603] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1e8db9-8c2a-4d57-a082-327c54baa21c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.681389] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b24fdb-7806-4307-8c9c-a279e8af653a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.781310] env[61824]: DEBUG nova.compute.manager [req-39c20811-d7fe-49ab-8ff5-87631ab6c02e req-895e3f5f-af8b-4f6e-8682-4e3667dae6c4 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Received event network-vif-plugged-1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1157.781495] env[61824]: DEBUG oslo_concurrency.lockutils [req-39c20811-d7fe-49ab-8ff5-87631ab6c02e req-895e3f5f-af8b-4f6e-8682-4e3667dae6c4 service nova] Acquiring lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.781659] env[61824]: DEBUG oslo_concurrency.lockutils [req-39c20811-d7fe-49ab-8ff5-87631ab6c02e req-895e3f5f-af8b-4f6e-8682-4e3667dae6c4 service nova] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.781793] env[61824]: DEBUG oslo_concurrency.lockutils [req-39c20811-d7fe-49ab-8ff5-87631ab6c02e req-895e3f5f-af8b-4f6e-8682-4e3667dae6c4 service nova] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.781977] env[61824]: DEBUG nova.compute.manager [req-39c20811-d7fe-49ab-8ff5-87631ab6c02e req-895e3f5f-af8b-4f6e-8682-4e3667dae6c4 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] No waiting events found dispatching network-vif-plugged-1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1157.782193] env[61824]: WARNING nova.compute.manager [req-39c20811-d7fe-49ab-8ff5-87631ab6c02e req-895e3f5f-af8b-4f6e-8682-4e3667dae6c4 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Received unexpected event network-vif-plugged-1cd7a243-c4f2-4cf5-9462-c13718ee7e00 for instance with vm_state building and task_state spawning. [ 1157.860776] env[61824]: DEBUG nova.network.neutron [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Successfully updated port: 1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1158.012957] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.032147] env[61824]: INFO nova.scheduler.client.report [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Deleted allocations for instance 4700de90-7230-43a5-87ae-f2c7ed9d4ea6 [ 1158.143249] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.143463] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.143601] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1158.364247] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.364247] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.364247] env[61824]: DEBUG nova.network.neutron [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1158.541329] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0dbce367-5549-4b58-ab7f-63cca1c6a799 tempest-ServerRescueNegativeTestJSON-1147293579 tempest-ServerRescueNegativeTestJSON-1147293579-project-member] Lock "4700de90-7230-43a5-87ae-f2c7ed9d4ea6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.978s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.894712] env[61824]: DEBUG nova.network.neutron [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1159.035400] env[61824]: DEBUG nova.network.neutron [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Updating instance_info_cache with network_info: [{"id": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "address": "fa:16:3e:f8:dd:98", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cd7a243-c4", "ovs_interfaceid": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.538165] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.538506] env[61824]: DEBUG nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Instance network_info: |[{"id": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "address": "fa:16:3e:f8:dd:98", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cd7a243-c4", "ovs_interfaceid": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1159.538961] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:dd:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1cd7a243-c4f2-4cf5-9462-c13718ee7e00', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1159.546964] env[61824]: DEBUG oslo.service.loopingcall [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.547213] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1159.547443] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f503cb31-a24c-46c3-a583-00d8a7502fa3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.567051] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1159.567051] env[61824]: value = "task-1276218" [ 1159.567051] env[61824]: _type = "Task" [ 1159.567051] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.574807] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276218, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.808320] env[61824]: DEBUG nova.compute.manager [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Received event network-changed-1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1159.808558] env[61824]: DEBUG nova.compute.manager [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Refreshing instance network info cache due to event network-changed-1cd7a243-c4f2-4cf5-9462-c13718ee7e00. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1159.808788] env[61824]: DEBUG oslo_concurrency.lockutils [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] Acquiring lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1159.808948] env[61824]: DEBUG oslo_concurrency.lockutils [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] Acquired lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.809436] env[61824]: DEBUG nova.network.neutron [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Refreshing network info cache for port 1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1160.078059] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276218, 'name': CreateVM_Task, 'duration_secs': 0.27069} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.078059] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1160.078662] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.078864] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.079251] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1160.079505] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3403223-1e5a-4d32-9af1-7a7ab43f03e6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.083811] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1160.083811] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]528cdb1b-11ec-9b3a-97c6-1f618033e8fe" [ 1160.083811] env[61824]: _type = "Task" [ 1160.083811] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.092233] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528cdb1b-11ec-9b3a-97c6-1f618033e8fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.152899] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Didn't find any instances for network info cache update. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1160.153156] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.153324] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.153497] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.153653] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.153808] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.153947] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_power_states {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.513650] env[61824]: DEBUG nova.network.neutron [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Updated VIF entry in instance network info cache for port 1cd7a243-c4f2-4cf5-9462-c13718ee7e00. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1160.514247] env[61824]: DEBUG nova.network.neutron [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Updating instance_info_cache with network_info: [{"id": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "address": "fa:16:3e:f8:dd:98", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cd7a243-c4", "ovs_interfaceid": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.593873] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]528cdb1b-11ec-9b3a-97c6-1f618033e8fe, 'name': SearchDatastore_Task, 'duration_secs': 0.012048} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.594189] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1160.594433] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1160.594669] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.594820] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.594996] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1160.595263] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ad0222f-7b5b-4689-8eec-9ad8fd69f7a4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.603054] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1160.603231] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1160.603905] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f25fa8e7-ac6a-4367-9297-e08204f1617a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.608929] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1160.608929] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]521ea6d2-0fa6-6b6c-aa09-9c249c81e2a2" [ 1160.608929] env[61824]: _type = "Task" [ 1160.608929] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.615790] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521ea6d2-0fa6-6b6c-aa09-9c249c81e2a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.657158] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Getting list of instances from cluster (obj){ [ 1160.657158] env[61824]: value = "domain-c8" [ 1160.657158] env[61824]: _type = "ClusterComputeResource" [ 1160.657158] env[61824]: } {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1160.658092] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b5fd73-c6ae-433b-8d1a-4a565e5bce1f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.668489] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Got total of 2 instances {{(pid=61824) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1160.668653] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid 8aa00861-bca6-4861-a5e5-0c538155c9ab {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1160.668852] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Triggering sync for uuid 0baaad18-8cbe-4b2d-872b-ee259534b1fb {{(pid=61824) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1160.669177] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.669385] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.669633] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.669814] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.669943] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1160.670602] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f835dc5-5e7a-422f-b430-fd9aa576a329 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.672956] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.017977] env[61824]: DEBUG oslo_concurrency.lockutils [req-8167bc08-5fb2-4e28-b169-5ee8c7f67279 req-45d19772-e79c-48b5-99e6-74ab650f367e service nova] Releasing lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.118618] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]521ea6d2-0fa6-6b6c-aa09-9c249c81e2a2, 'name': SearchDatastore_Task, 'duration_secs': 0.007941} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.119433] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98b2573b-354e-4bc8-9a9b-c56d90c92c63 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.124211] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1161.124211] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]526e21fd-dfc4-800e-8308-c11700839a90" [ 1161.124211] env[61824]: _type = "Task" [ 1161.124211] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.131161] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526e21fd-dfc4-800e-8308-c11700839a90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.176117] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.176329] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.176489] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.176944] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1161.177367] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e197a2-ac51-421a-a522-70761acc663e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.179774] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.185225] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36c60e7-3d52-4711-8f1e-8bc95899b9c5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.198679] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8a8274-1e9b-4a59-aeb0-27e198b36314 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.205054] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5368c32f-4aaa-4539-b947-5df8f62ba79d {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.234285] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180595MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1161.234419] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.234608] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.634495] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]526e21fd-dfc4-800e-8308-c11700839a90, 'name': SearchDatastore_Task, 'duration_secs': 0.008669} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.634773] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.635057] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 0baaad18-8cbe-4b2d-872b-ee259534b1fb/0baaad18-8cbe-4b2d-872b-ee259534b1fb.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1161.635328] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c8d27d9-33a4-4eea-98d9-69868ddcf394 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.641093] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1161.641093] env[61824]: value = "task-1276219" [ 1161.641093] env[61824]: _type = "Task" [ 1161.641093] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.648187] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.150831] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276219, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.399529} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.151242] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] 0baaad18-8cbe-4b2d-872b-ee259534b1fb/0baaad18-8cbe-4b2d-872b-ee259534b1fb.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1162.151402] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1162.151649] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03fdb597-e270-4abe-95ef-22ebdad527f0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.158113] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1162.158113] env[61824]: value = "task-1276220" [ 1162.158113] env[61824]: _type = "Task" [ 1162.158113] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.165441] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276220, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.258684] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8aa00861-bca6-4861-a5e5-0c538155c9ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.258840] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 0baaad18-8cbe-4b2d-872b-ee259534b1fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1162.259205] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1162.259205] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1162.291537] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0530febf-042b-4df3-9d62-c1cdb14cb3f8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.298835] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edaa5bfe-3c14-4604-a62b-43372f2ec53b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.328670] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ed31ee-9289-4452-b81d-1681db23ea4e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.335274] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecc3426-bfac-4af3-94ae-abf6f580f5f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.347710] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.667308] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276220, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056246} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.667629] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1162.668381] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d50f0a3-d454-4640-a61d-de5d4b5233b3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.689407] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 0baaad18-8cbe-4b2d-872b-ee259534b1fb/0baaad18-8cbe-4b2d-872b-ee259534b1fb.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1162.689632] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea58fb28-b035-41bd-9f47-2ba45e445086 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.707672] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1162.707672] env[61824]: value = "task-1276221" [ 1162.707672] env[61824]: _type = "Task" [ 1162.707672] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.714777] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276221, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.850737] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1163.218077] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276221, 'name': ReconfigVM_Task, 'duration_secs': 0.240673} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.218448] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 0baaad18-8cbe-4b2d-872b-ee259534b1fb/0baaad18-8cbe-4b2d-872b-ee259534b1fb.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1163.218968] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6593c9b1-febb-4b87-8c06-0f78d6f5199e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.225198] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1163.225198] env[61824]: value = "task-1276222" [ 1163.225198] env[61824]: _type = "Task" [ 1163.225198] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.232248] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276222, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.355427] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1163.355669] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.121s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.734739] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276222, 'name': Rename_Task, 'duration_secs': 0.153916} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.735016] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1163.735266] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef2717c9-a3b2-437b-b85b-3b05a5483eb7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.742310] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1163.742310] env[61824]: value = "task-1276223" [ 1163.742310] env[61824]: _type = "Task" [ 1163.742310] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.749450] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276223, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.251775] env[61824]: DEBUG oslo_vmware.api [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276223, 'name': PowerOnVM_Task, 'duration_secs': 0.40702} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.252178] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1164.252289] env[61824]: INFO nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Took 6.61 seconds to spawn the instance on the hypervisor. [ 1164.252474] env[61824]: DEBUG nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1164.253232] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce633f7e-590d-4694-8cf7-70333d1d50ce {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.769964] env[61824]: INFO nova.compute.manager [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Took 11.28 seconds to build instance. [ 1165.271775] env[61824]: DEBUG oslo_concurrency.lockutils [None req-e3a6a866-4b0d-4993-9a75-790998b0baae tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.793s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.272106] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.602s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.272301] env[61824]: INFO nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] During sync_power_state the instance has a pending task (spawning). Skip. [ 1165.272490] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.380730] env[61824]: DEBUG nova.compute.manager [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Received event network-changed-1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1165.380978] env[61824]: DEBUG nova.compute.manager [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Refreshing instance network info cache due to event network-changed-1cd7a243-c4f2-4cf5-9462-c13718ee7e00. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1165.381249] env[61824]: DEBUG oslo_concurrency.lockutils [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] Acquiring lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.381404] env[61824]: DEBUG oslo_concurrency.lockutils [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] Acquired lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.381570] env[61824]: DEBUG nova.network.neutron [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Refreshing network info cache for port 1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1166.088621] env[61824]: DEBUG nova.network.neutron [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Updated VIF entry in instance network info cache for port 1cd7a243-c4f2-4cf5-9462-c13718ee7e00. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1166.089098] env[61824]: DEBUG nova.network.neutron [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Updating instance_info_cache with network_info: [{"id": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "address": "fa:16:3e:f8:dd:98", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1cd7a243-c4", "ovs_interfaceid": "1cd7a243-c4f2-4cf5-9462-c13718ee7e00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.591979] env[61824]: DEBUG oslo_concurrency.lockutils [req-754b0c42-3e0e-43d2-896b-ed217e98686e req-cdd021e9-f0a7-4dcb-95de-cf72c23d1e49 service nova] Releasing lock "refresh_cache-0baaad18-8cbe-4b2d-872b-ee259534b1fb" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.719065] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1184.719350] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.222647] env[61824]: DEBUG nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1185.742402] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.742697] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.744540] env[61824]: INFO nova.compute.claims [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1186.803268] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2f6e77-75a6-4ccc-93bb-23e595749e19 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.811529] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60981442-eb1e-493d-9e1f-637770a14814 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.842458] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a2b244-f4f2-4bbc-9f2c-fef7eab64297 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.850168] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f341997-258f-4411-81f5-d7e6d80166a0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.863486] env[61824]: DEBUG nova.compute.provider_tree [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.366535] env[61824]: DEBUG nova.scheduler.client.report [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1187.871940] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.129s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.872403] env[61824]: DEBUG nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1188.377624] env[61824]: DEBUG nova.compute.utils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1188.379634] env[61824]: DEBUG nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1188.379634] env[61824]: DEBUG nova.network.neutron [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1188.432689] env[61824]: DEBUG nova.policy [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad4e8224518048e583d27b94721b3c22', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd21256931aa54010b72beca4fb798f63', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1188.688035] env[61824]: DEBUG nova.network.neutron [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Successfully created port: e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1188.883129] env[61824]: DEBUG nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1189.388591] env[61824]: INFO nova.virt.block_device [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Booting with volume 65021a4c-756d-497a-9bcc-92dfce6a2da2 at /dev/sda [ 1189.425151] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53c21e9a-62ec-444e-b84e-f97eca26f927 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.436264] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f714b4d-7b20-47f9-b73a-3c466dd802b3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.466572] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-593fdfc6-b249-475b-87bf-6f21d7f9512c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.475399] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279a3c33-7a45-49b1-ad8b-fa609ab81fc0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.499951] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fd4f19-d12d-4542-ad1e-1e25988fddb8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.506783] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-266b1d5e-e485-43c1-bf83-10b6148a131e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.521317] env[61824]: DEBUG nova.virt.block_device [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating existing volume attachment record: 84b659fd-efb4-4627-b515-339d14e6b44d {{(pid=61824) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1190.044484] env[61824]: DEBUG nova.compute.manager [req-de7c02c9-5c0f-4b99-930f-a9281cd29e0f req-798ff6f0-8acc-44ad-aaa2-3b7c973a4787 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Received event network-vif-plugged-e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1190.044726] env[61824]: DEBUG oslo_concurrency.lockutils [req-de7c02c9-5c0f-4b99-930f-a9281cd29e0f req-798ff6f0-8acc-44ad-aaa2-3b7c973a4787 service nova] Acquiring lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1190.044926] env[61824]: DEBUG oslo_concurrency.lockutils [req-de7c02c9-5c0f-4b99-930f-a9281cd29e0f req-798ff6f0-8acc-44ad-aaa2-3b7c973a4787 service nova] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1190.045087] env[61824]: DEBUG oslo_concurrency.lockutils [req-de7c02c9-5c0f-4b99-930f-a9281cd29e0f req-798ff6f0-8acc-44ad-aaa2-3b7c973a4787 service nova] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.045259] env[61824]: DEBUG nova.compute.manager [req-de7c02c9-5c0f-4b99-930f-a9281cd29e0f req-798ff6f0-8acc-44ad-aaa2-3b7c973a4787 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] No waiting events found dispatching network-vif-plugged-e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1190.045424] env[61824]: WARNING nova.compute.manager [req-de7c02c9-5c0f-4b99-930f-a9281cd29e0f req-798ff6f0-8acc-44ad-aaa2-3b7c973a4787 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Received unexpected event network-vif-plugged-e8ae94cc-31af-4110-b28d-d957b637cf0e for instance with vm_state building and task_state block_device_mapping. [ 1190.125628] env[61824]: DEBUG nova.network.neutron [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Successfully updated port: e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1190.627615] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.627778] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.627934] env[61824]: DEBUG nova.network.neutron [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1191.159537] env[61824]: DEBUG nova.network.neutron [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1191.284710] env[61824]: DEBUG nova.network.neutron [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [{"id": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "address": "fa:16:3e:e3:4d:2f", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8ae94cc-31", "ovs_interfaceid": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.600737] env[61824]: DEBUG nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1191.601297] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1191.601545] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1191.601711] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1191.601896] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1191.602057] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1191.602210] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1191.602425] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1191.602624] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1191.602816] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1191.602984] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1191.603177] env[61824]: DEBUG nova.virt.hardware [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1191.604027] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4780874-0576-4f76-bff4-f8e81d04bf5b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.612676] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b2e193-806b-465c-8401-cd94d476d4d7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.787955] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.788321] env[61824]: DEBUG nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Instance network_info: |[{"id": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "address": "fa:16:3e:e3:4d:2f", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8ae94cc-31", "ovs_interfaceid": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1191.788754] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:4d:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd6fab536-1e48-4d07-992a-076f0e6d089c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8ae94cc-31af-4110-b28d-d957b637cf0e', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1191.796112] env[61824]: DEBUG oslo.service.loopingcall [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.796323] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1191.796545] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd519c4c-5468-4b92-bf1d-2e45838ecb12 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.818141] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1191.818141] env[61824]: value = "task-1276232" [ 1191.818141] env[61824]: _type = "Task" [ 1191.818141] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.826126] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276232, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.070239] env[61824]: DEBUG nova.compute.manager [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Received event network-changed-e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1192.070457] env[61824]: DEBUG nova.compute.manager [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Refreshing instance network info cache due to event network-changed-e8ae94cc-31af-4110-b28d-d957b637cf0e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1192.070705] env[61824]: DEBUG oslo_concurrency.lockutils [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] Acquiring lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.070856] env[61824]: DEBUG oslo_concurrency.lockutils [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] Acquired lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.071035] env[61824]: DEBUG nova.network.neutron [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Refreshing network info cache for port e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1192.328974] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276232, 'name': CreateVM_Task, 'duration_secs': 0.325413} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.329294] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1192.329845] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'disk_bus': None, 'boot_index': 0, 'guest_format': None, 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274253', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'name': 'volume-65021a4c-756d-497a-9bcc-92dfce6a2da2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76', 'attached_at': '', 'detached_at': '', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'serial': '65021a4c-756d-497a-9bcc-92dfce6a2da2'}, 'mount_device': '/dev/sda', 'attachment_id': '84b659fd-efb4-4627-b515-339d14e6b44d', 'device_type': None, 'volume_type': None}], 'swap': None} {{(pid=61824) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1192.330076] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Root volume attach. Driver type: vmdk {{(pid=61824) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1192.330842] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b09425-8d43-4be0-a8b3-e886fe9c4b29 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.339236] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a0e645-b025-489f-95c7-a44814ad8f11 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.345086] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba8ac92-ae2a-4ea7-846d-602e7999aaed {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.350534] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-d4a9e2d4-bb7a-48db-9f81-51266ec70934 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.357954] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1192.357954] env[61824]: value = "task-1276233" [ 1192.357954] env[61824]: _type = "Task" [ 1192.357954] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.365100] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276233, 'name': RelocateVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.765575] env[61824]: DEBUG nova.network.neutron [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updated VIF entry in instance network info cache for port e8ae94cc-31af-4110-b28d-d957b637cf0e. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1192.765959] env[61824]: DEBUG nova.network.neutron [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [{"id": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "address": "fa:16:3e:e3:4d:2f", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8ae94cc-31", "ovs_interfaceid": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.867591] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276233, 'name': RelocateVM_Task, 'duration_secs': 0.401107} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1192.867822] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Volume attach. Driver type: vmdk {{(pid=61824) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1192.868036] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274253', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'name': 'volume-65021a4c-756d-497a-9bcc-92dfce6a2da2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76', 'attached_at': '', 'detached_at': '', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'serial': '65021a4c-756d-497a-9bcc-92dfce6a2da2'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1192.868787] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8998d74d-8977-4521-80df-37f07664c108 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.883584] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11de1a2c-10be-4171-a9e5-4e590070a707 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.904741] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-65021a4c-756d-497a-9bcc-92dfce6a2da2/volume-65021a4c-756d-497a-9bcc-92dfce6a2da2.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1192.905272] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39df8517-297a-4812-9b56-8a0089a8bf2e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.924772] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1192.924772] env[61824]: value = "task-1276234" [ 1192.924772] env[61824]: _type = "Task" [ 1192.924772] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.932231] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276234, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.268577] env[61824]: DEBUG oslo_concurrency.lockutils [req-e5b04080-cb78-45e6-ad9d-57a792a97dad req-dbbb9dd3-dee8-4426-a78d-6fa863b94dbd service nova] Releasing lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1193.434580] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276234, 'name': ReconfigVM_Task, 'duration_secs': 0.287357} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.434925] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-65021a4c-756d-497a-9bcc-92dfce6a2da2/volume-65021a4c-756d-497a-9bcc-92dfce6a2da2.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1193.439438] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aed245a2-f1de-4809-9e22-908773a6af78 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.453975] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1193.453975] env[61824]: value = "task-1276235" [ 1193.453975] env[61824]: _type = "Task" [ 1193.453975] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.462859] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276235, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.964282] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276235, 'name': ReconfigVM_Task, 'duration_secs': 0.136015} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.964624] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274253', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'name': 'volume-65021a4c-756d-497a-9bcc-92dfce6a2da2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76', 'attached_at': '', 'detached_at': '', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'serial': '65021a4c-756d-497a-9bcc-92dfce6a2da2'} {{(pid=61824) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1193.965147] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fc63f134-2200-46e6-b2bf-3e37dbb279d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.972241] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1193.972241] env[61824]: value = "task-1276236" [ 1193.972241] env[61824]: _type = "Task" [ 1193.972241] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.979837] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276236, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.483171] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276236, 'name': Rename_Task, 'duration_secs': 0.131721} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.483517] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1194.483685] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cb6991b-839d-43e5-8659-324d52a5966c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.490234] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1194.490234] env[61824]: value = "task-1276237" [ 1194.490234] env[61824]: _type = "Task" [ 1194.490234] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.497224] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.999837] env[61824]: DEBUG oslo_vmware.api [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276237, 'name': PowerOnVM_Task, 'duration_secs': 0.460307} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.000126] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1195.000319] env[61824]: INFO nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Took 3.40 seconds to spawn the instance on the hypervisor. [ 1195.000505] env[61824]: DEBUG nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1195.001261] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341cf3aa-e0fd-46e7-a0f9-db896e0821c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.518874] env[61824]: INFO nova.compute.manager [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Took 9.79 seconds to build instance. [ 1195.876098] env[61824]: DEBUG nova.compute.manager [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Received event network-changed-87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1195.876329] env[61824]: DEBUG nova.compute.manager [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Refreshing instance network info cache due to event network-changed-87ec4b1c-0d02-46ba-878d-b392ae74f472. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1195.876552] env[61824]: DEBUG oslo_concurrency.lockutils [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] Acquiring lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1195.876702] env[61824]: DEBUG oslo_concurrency.lockutils [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] Acquired lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.876868] env[61824]: DEBUG nova.network.neutron [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Refreshing network info cache for port 87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1196.020953] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0037bcb7-d169-44a5-852f-8424f9836750 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.301s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.581373] env[61824]: DEBUG nova.network.neutron [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updated VIF entry in instance network info cache for port 87ec4b1c-0d02-46ba-878d-b392ae74f472. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1196.581888] env[61824]: DEBUG nova.network.neutron [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [{"id": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "address": "fa:16:3e:83:f8:c1", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ec4b1c-0d", "ovs_interfaceid": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.085180] env[61824]: DEBUG oslo_concurrency.lockutils [req-424972f2-f4fb-4167-96f9-f4b1a8e726a2 req-163e1863-ac1c-4770-a523-e6c330c6908f service nova] Releasing lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1197.369877] env[61824]: DEBUG nova.compute.manager [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Stashing vm_state: active {{(pid=61824) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1197.892134] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.892409] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1197.903314] env[61824]: DEBUG nova.compute.manager [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Received event network-changed-e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1197.903521] env[61824]: DEBUG nova.compute.manager [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Refreshing instance network info cache due to event network-changed-e8ae94cc-31af-4110-b28d-d957b637cf0e. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1197.903804] env[61824]: DEBUG oslo_concurrency.lockutils [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] Acquiring lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.903966] env[61824]: DEBUG oslo_concurrency.lockutils [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] Acquired lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.904150] env[61824]: DEBUG nova.network.neutron [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Refreshing network info cache for port e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1198.398145] env[61824]: INFO nova.compute.claims [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1198.649184] env[61824]: DEBUG nova.network.neutron [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updated VIF entry in instance network info cache for port e8ae94cc-31af-4110-b28d-d957b637cf0e. {{(pid=61824) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1198.649562] env[61824]: DEBUG nova.network.neutron [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [{"id": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "address": "fa:16:3e:e3:4d:2f", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8ae94cc-31", "ovs_interfaceid": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.904729] env[61824]: INFO nova.compute.resource_tracker [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating resource usage from migration 6e21da93-c1b9-4d6d-8923-35747bfa648c [ 1198.963371] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dede122d-0c1f-4ec3-b62e-79a6949630bc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.971561] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80feaa4-a43b-4435-9712-adafb1028e33 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.000708] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e04a5d-7813-45c9-a2f7-3f185ac01a98 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.007759] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f237e0a2-a369-445f-8dd1-67afebb1c82b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.022092] env[61824]: DEBUG nova.compute.provider_tree [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1199.152145] env[61824]: DEBUG oslo_concurrency.lockutils [req-8462b8c0-69a2-4a71-9810-748040d427ef req-67ec849d-55ed-42c1-b025-521e77ee3894 service nova] Releasing lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.525435] env[61824]: DEBUG nova.scheduler.client.report [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1200.029846] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.137s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1200.030221] env[61824]: INFO nova.compute.manager [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Migrating [ 1200.277641] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.277935] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.544315] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1200.544559] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.544688] env[61824]: DEBUG nova.network.neutron [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1200.782905] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.783171] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1200.783264] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Rebuilding the list of instances to heal {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1201.251242] env[61824]: DEBUG nova.network.neutron [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [{"id": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "address": "fa:16:3e:e3:4d:2f", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8ae94cc-31", "ovs_interfaceid": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.310868] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1201.311039] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquired lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1201.311192] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Forcefully refreshing network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1201.311348] env[61824]: DEBUG nova.objects.instance [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lazy-loading 'info_cache' on Instance uuid 8aa00861-bca6-4861-a5e5-0c538155c9ab {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.754088] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.058959] env[61824]: DEBUG nova.network.neutron [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [{"id": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "address": "fa:16:3e:83:f8:c1", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87ec4b1c-0d", "ovs_interfaceid": "87ec4b1c-0d02-46ba-878d-b392ae74f472", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.270454] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda7d4d5-2412-420d-92fd-80e44897abef {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.289286] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76' progress to 0 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1203.562204] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Releasing lock "refresh_cache-8aa00861-bca6-4861-a5e5-0c538155c9ab" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1203.562410] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updated the network info_cache for instance {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1203.562618] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.562780] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.562928] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.563089] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.563236] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.563380] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.563506] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1203.563647] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1203.795507] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1203.795822] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-017e5ddd-9698-4e7b-af67-e61a58ab5448 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.804450] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1203.804450] env[61824]: value = "task-1276238" [ 1203.804450] env[61824]: _type = "Task" [ 1203.804450] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.812620] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.066533] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.066835] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.066914] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1204.067086] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1204.067985] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad49f78d-aa55-403f-907c-a1945c0b56ab {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.075554] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fc7a6d-5b7d-4979-9dd3-6823fb0df350 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.088969] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdbd84e-b3a4-410b-b629-47f1504a704e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.094866] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b435d4-3434-44e6-a6ec-8287be48c6dd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.123923] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180595MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1204.124066] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.124251] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.313938] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276238, 'name': PowerOffVM_Task, 'duration_secs': 0.237209} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.314239] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1204.314430] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76' progress to 17 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1204.820970] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1204.821237] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1204.821281] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1204.821450] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1204.821647] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1204.821814] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1204.822032] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1204.822205] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1204.822376] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1204.822543] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1204.822741] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1204.827735] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7199ac6-e4bd-4df7-8b47-37c2cf8e2c14 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.843587] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1204.843587] env[61824]: value = "task-1276239" [ 1204.843587] env[61824]: _type = "Task" [ 1204.843587] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.851261] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.132076] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Applying migration context for instance 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76 as it has an incoming, in-progress migration 6e21da93-c1b9-4d6d-8923-35747bfa648c. Migration status is migrating {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1205.132841] env[61824]: INFO nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating resource usage from migration 6e21da93-c1b9-4d6d-8923-35747bfa648c [ 1205.150882] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8aa00861-bca6-4861-a5e5-0c538155c9ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1205.151054] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 0baaad18-8cbe-4b2d-872b-ee259534b1fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1205.151184] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Migration 6e21da93-c1b9-4d6d-8923-35747bfa648c is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1205.151304] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1205.151488] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1205.151727] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1205.167884] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Refreshing inventories for resource provider 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1205.181791] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Updating ProviderTree inventory for provider 8d9ece0e-729c-4d69-b741-d89969200860 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1205.181980] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Updating inventory in ProviderTree for provider 8d9ece0e-729c-4d69-b741-d89969200860 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1205.193049] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Refreshing aggregate associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, aggregates: None {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1205.210106] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Refreshing trait associations for resource provider 8d9ece0e-729c-4d69-b741-d89969200860, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61824) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1205.262399] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20585678-dc3e-4a8d-a8d6-70687cec3bf3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.269814] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88553cba-6abc-428e-a390-a3547f0e1429 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.298971] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5728b3-c351-42b6-809d-580ae6bc0cb4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.305726] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e88e4ae-c919-4c30-9c07-edcb61567af3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.319337] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.352330] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276239, 'name': ReconfigVM_Task, 'duration_secs': 0.143819} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.352618] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76' progress to 33 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1205.822492] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1205.858408] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1205.858689] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1205.858820] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1205.859015] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1205.859177] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1205.859328] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1205.859546] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1205.859730] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1205.859908] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1205.860099] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1205.860284] env[61824]: DEBUG nova.virt.hardware [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1205.867036] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfiguring VM instance instance-0000006f to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1205.867036] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63b9c5cb-0291-4660-a628-04921edbee40 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.884776] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1205.884776] env[61824]: value = "task-1276240" [ 1205.884776] env[61824]: _type = "Task" [ 1205.884776] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.892342] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276240, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.327561] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1206.327884] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.203s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.394801] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276240, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.895680] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276240, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.396279] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276240, 'name': ReconfigVM_Task, 'duration_secs': 1.142935} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.396676] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfigured VM instance instance-0000006f to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1207.397353] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b9cd32-3cb2-453b-90be-39e814fb263a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.418522] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-65021a4c-756d-497a-9bcc-92dfce6a2da2/volume-65021a4c-756d-497a-9bcc-92dfce6a2da2.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1207.418770] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cca68632-3890-4ef1-9886-4505b3416317 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.445371] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1207.445371] env[61824]: value = "task-1276241" [ 1207.445371] env[61824]: _type = "Task" [ 1207.445371] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.452896] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276241, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.954325] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276241, 'name': ReconfigVM_Task, 'duration_secs': 0.234822} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.954675] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-65021a4c-756d-497a-9bcc-92dfce6a2da2/volume-65021a4c-756d-497a-9bcc-92dfce6a2da2.vmdk or device None with type thin {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1207.954966] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76' progress to 50 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1208.461919] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-335d0d42-31a6-4dd8-a4d2-5d36969583f1 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.482179] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065b28e4-5b06-4278-88df-c8c482c86687 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.499010] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76' progress to 67 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1210.141981] env[61824]: DEBUG nova.network.neutron [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Port e8ae94cc-31af-4110-b28d-d957b637cf0e binding to destination host cpu-1 is already ACTIVE {{(pid=61824) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1211.162374] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.162840] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1211.162840] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.198051] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1212.198322] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1212.198443] env[61824]: DEBUG nova.network.neutron [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1212.895609] env[61824]: DEBUG nova.network.neutron [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [{"id": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "address": "fa:16:3e:e3:4d:2f", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8ae94cc-31", "ovs_interfaceid": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.398575] env[61824]: DEBUG oslo_concurrency.lockutils [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1213.909613] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f86f04-90b2-4f79-a743-1a06136beb07 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.916853] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef817d0-32f6-4da1-a439-3a5a3ff40748 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.011953] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfda7cb-cd6e-43d5-91db-fa4375f5ee3a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.030878] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076670b9-b1c1-4637-abfc-f3351a3ee4ec {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.037130] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76' progress to 83 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1215.542918] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1215.543274] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-15d81371-efe0-452d-9a66-cd062e094800 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.551429] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1215.551429] env[61824]: value = "task-1276242" [ 1215.551429] env[61824]: _type = "Task" [ 1215.551429] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.558941] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.063741] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276242, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.562074] env[61824]: DEBUG oslo_vmware.api [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276242, 'name': PowerOnVM_Task, 'duration_secs': 0.524739} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.562344] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1216.562530] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-59ee268e-1b9d-4d3d-971e-f905d6f8335b tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76' progress to 100 {{(pid=61824) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1219.081607] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.081939] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.082158] env[61824]: DEBUG nova.compute.manager [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Going to confirm migration 4 {{(pid=61824) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1219.617082] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1219.617248] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquired lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.617429] env[61824]: DEBUG nova.network.neutron [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1219.617611] env[61824]: DEBUG nova.objects.instance [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'info_cache' on Instance uuid 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1220.851639] env[61824]: DEBUG nova.network.neutron [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [{"id": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "address": "fa:16:3e:e3:4d:2f", "network": {"id": "8880fda9-f0ec-4620-a895-da0e5fb916a9", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-777192001-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d21256931aa54010b72beca4fb798f63", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d6fab536-1e48-4d07-992a-076f0e6d089c", "external-id": "nsx-vlan-transportzone-61", "segmentation_id": 61, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8ae94cc-31", "ovs_interfaceid": "e8ae94cc-31af-4110-b28d-d957b637cf0e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.354539] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Releasing lock "refresh_cache-6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.354801] env[61824]: DEBUG nova.objects.instance [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'migration_context' on Instance uuid 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76 {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1221.857738] env[61824]: DEBUG nova.objects.base [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Object Instance<6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76> lazy-loaded attributes: info_cache,migration_context {{(pid=61824) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1221.858708] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f08147-dffa-477c-a001-5ec472eb6f44 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.878248] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0c548e3-ce58-4269-ae59-d041d7776f82 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.883065] env[61824]: DEBUG oslo_vmware.api [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1221.883065] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52dec203-5757-16f4-9cca-cdab4e735388" [ 1221.883065] env[61824]: _type = "Task" [ 1221.883065] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.890418] env[61824]: DEBUG oslo_vmware.api [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52dec203-5757-16f4-9cca-cdab4e735388, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.393738] env[61824]: DEBUG oslo_vmware.api [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52dec203-5757-16f4-9cca-cdab4e735388, 'name': SearchDatastore_Task, 'duration_secs': 0.006717} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.394040] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.394323] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.956650] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bb41dd-abad-428d-bb34-65c7c02d5fdd {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.964565] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67d18d6-3205-4487-a241-2acc70394635 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.993414] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b85bfbb-e292-4bdd-bf60-dcec44453017 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.000353] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04fe7987-31eb-4f58-be44-8826731ff93a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.013548] env[61824]: DEBUG nova.compute.provider_tree [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1223.517104] env[61824]: DEBUG nova.scheduler.client.report [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1224.529060] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.134s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1225.085843] env[61824]: INFO nova.scheduler.client.report [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted allocation for migration 6e21da93-c1b9-4d6d-8923-35747bfa648c [ 1225.454101] env[61824]: INFO nova.compute.manager [None req-ffcc4913-c4d1-43f5-bf35-dd3a9bcf5818 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Get console output [ 1225.454483] env[61824]: WARNING nova.virt.vmwareapi.driver [None req-ffcc4913-c4d1-43f5-bf35-dd3a9bcf5818 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] The console log is missing. Check your VSPC configuration [ 1225.592418] env[61824]: DEBUG oslo_concurrency.lockutils [None req-0b8a3cf4-8a85-4d56-81b3-8edbd1bc86fa tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.510s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.320368] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.320789] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.320883] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1252.321074] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.321250] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.324365] env[61824]: INFO nova.compute.manager [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Terminating instance [ 1252.326259] env[61824]: DEBUG nova.compute.manager [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1252.326478] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1252.326717] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f1dba74-8a3c-49e3-89c1-0589420c711a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.333904] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1252.333904] env[61824]: value = "task-1276243" [ 1252.333904] env[61824]: _type = "Task" [ 1252.333904] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.341739] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.843961] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276243, 'name': PowerOffVM_Task, 'duration_secs': 0.15776} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.844274] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1252.844471] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Volume detach. Driver type: vmdk {{(pid=61824) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1252.844667] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274253', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'name': 'volume-65021a4c-756d-497a-9bcc-92dfce6a2da2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76', 'attached_at': '2024-10-12T11:04:28.000000', 'detached_at': '', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'serial': '65021a4c-756d-497a-9bcc-92dfce6a2da2'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1252.845427] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d6e13e-0b14-4e4a-9398-690c7276d70f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.863634] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71043ff-1f18-4b4d-8a78-310dcc46fbdb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.869851] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4924fd02-501d-4700-bc8c-e3fdb949f62f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.887591] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b3af44-a631-4c81-8adc-a72723d7f5f3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.901407] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] The volume has not been displaced from its original location: [datastore2] volume-65021a4c-756d-497a-9bcc-92dfce6a2da2/volume-65021a4c-756d-497a-9bcc-92dfce6a2da2.vmdk. No consolidation needed. {{(pid=61824) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1252.906608] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfiguring VM instance instance-0000006f to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1252.906854] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ae0f3fd-08a3-46cf-aa8b-4d5bbe20a08a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.923521] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1252.923521] env[61824]: value = "task-1276244" [ 1252.923521] env[61824]: _type = "Task" [ 1252.923521] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.930629] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276244, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.433123] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276244, 'name': ReconfigVM_Task, 'duration_secs': 0.142453} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.433506] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Reconfigured VM instance instance-0000006f to detach disk 2000 {{(pid=61824) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1253.438056] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d1eef6d-4950-455e-9b32-bb546e585e62 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.452605] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1253.452605] env[61824]: value = "task-1276245" [ 1253.452605] env[61824]: _type = "Task" [ 1253.452605] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.459876] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276245, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.964233] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276245, 'name': ReconfigVM_Task, 'duration_secs': 0.119573} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.964527] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-274253', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'name': 'volume-65021a4c-756d-497a-9bcc-92dfce6a2da2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76', 'attached_at': '2024-10-12T11:04:28.000000', 'detached_at': '', 'volume_id': '65021a4c-756d-497a-9bcc-92dfce6a2da2', 'serial': '65021a4c-756d-497a-9bcc-92dfce6a2da2'} {{(pid=61824) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1253.964815] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1253.965581] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217d899b-1687-4ec5-8707-c387d08cae50 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.971720] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1253.971938] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c83f65a8-be8f-4558-ab28-2e74bd2c4c4b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.038464] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1254.038706] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1254.038898] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleting the datastore file [datastore2] 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76 {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1254.039225] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9aa9e13-898e-4b87-94a3-d03d9bcbe79f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.047182] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1254.047182] env[61824]: value = "task-1276247" [ 1254.047182] env[61824]: _type = "Task" [ 1254.047182] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.054937] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276247, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.557515] env[61824]: DEBUG oslo_vmware.api [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276247, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.081418} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.557917] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1254.557962] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1254.558174] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1254.558362] env[61824]: INFO nova.compute.manager [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1254.558603] env[61824]: DEBUG oslo.service.loopingcall [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1254.558797] env[61824]: DEBUG nova.compute.manager [-] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1254.558891] env[61824]: DEBUG nova.network.neutron [-] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1254.961269] env[61824]: DEBUG nova.compute.manager [req-1b6457e9-0d27-4493-a995-e0969514642d req-ad91319a-6650-49f5-8606-319b77998b63 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Received event network-vif-deleted-e8ae94cc-31af-4110-b28d-d957b637cf0e {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1254.961516] env[61824]: INFO nova.compute.manager [req-1b6457e9-0d27-4493-a995-e0969514642d req-ad91319a-6650-49f5-8606-319b77998b63 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Neutron deleted interface e8ae94cc-31af-4110-b28d-d957b637cf0e; detaching it from the instance and deleting it from the info cache [ 1254.961744] env[61824]: DEBUG nova.network.neutron [req-1b6457e9-0d27-4493-a995-e0969514642d req-ad91319a-6650-49f5-8606-319b77998b63 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.438059] env[61824]: DEBUG nova.network.neutron [-] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.465029] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7213967d-21f0-433a-bfd5-b2f735ca29fa {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.475112] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c02f3e-9473-4d1e-88cb-8edae219abf2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.499837] env[61824]: DEBUG nova.compute.manager [req-1b6457e9-0d27-4493-a995-e0969514642d req-ad91319a-6650-49f5-8606-319b77998b63 service nova] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Detach interface failed, port_id=e8ae94cc-31af-4110-b28d-d957b637cf0e, reason: Instance 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76 could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1255.947334] env[61824]: INFO nova.compute.manager [-] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Took 1.39 seconds to deallocate network for instance. [ 1256.492550] env[61824]: INFO nova.compute.manager [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Took 0.55 seconds to detach 1 volumes for instance. [ 1256.494738] env[61824]: DEBUG nova.compute.manager [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76] Deleting volume: 65021a4c-756d-497a-9bcc-92dfce6a2da2 {{(pid=61824) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1257.031443] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.031798] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.031899] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.055116] env[61824]: INFO nova.scheduler.client.report [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted allocations for instance 6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76 [ 1257.562399] env[61824]: DEBUG oslo_concurrency.lockutils [None req-51584f1a-39d7-4b33-a553-25e79d8f5de7 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "6c8eb3cf-5d51-4fe5-8c7d-bacad8a64a76" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.242s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.859654] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.859875] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.860112] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1257.860310] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1257.860490] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1257.862716] env[61824]: INFO nova.compute.manager [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Terminating instance [ 1257.864454] env[61824]: DEBUG nova.compute.manager [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1257.864655] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1257.865502] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3235c1ca-899d-4b08-8951-f13f2e2f025e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.873651] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1257.873873] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9d601ac-4738-4625-8764-0bc36c081225 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.880187] env[61824]: DEBUG oslo_vmware.api [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1257.880187] env[61824]: value = "task-1276249" [ 1257.880187] env[61824]: _type = "Task" [ 1257.880187] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.887751] env[61824]: DEBUG oslo_vmware.api [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276249, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.390572] env[61824]: DEBUG oslo_vmware.api [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276249, 'name': PowerOffVM_Task, 'duration_secs': 0.169814} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.391037] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1258.391037] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1258.391254] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc9b9b00-d4e4-4d51-86e1-8007f190a52b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.450189] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1258.450427] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1258.450601] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleting the datastore file [datastore2] 0baaad18-8cbe-4b2d-872b-ee259534b1fb {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1258.450878] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-772d13e5-edf9-42be-b89f-2a817965a8b5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.456429] env[61824]: DEBUG oslo_vmware.api [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1258.456429] env[61824]: value = "task-1276251" [ 1258.456429] env[61824]: _type = "Task" [ 1258.456429] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.463534] env[61824]: DEBUG oslo_vmware.api [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.966826] env[61824]: DEBUG oslo_vmware.api [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126578} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.967080] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1258.967274] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1258.967455] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1258.967631] env[61824]: INFO nova.compute.manager [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1258.967868] env[61824]: DEBUG oslo.service.loopingcall [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1258.968077] env[61824]: DEBUG nova.compute.manager [-] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1258.968175] env[61824]: DEBUG nova.network.neutron [-] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1259.188175] env[61824]: DEBUG nova.compute.manager [req-4d6b7e79-922d-49d0-9e7d-17a8d52c2953 req-eafa435c-8019-47e2-b7e5-5c2b64abbce7 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Received event network-vif-deleted-1cd7a243-c4f2-4cf5-9462-c13718ee7e00 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1259.188349] env[61824]: INFO nova.compute.manager [req-4d6b7e79-922d-49d0-9e7d-17a8d52c2953 req-eafa435c-8019-47e2-b7e5-5c2b64abbce7 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Neutron deleted interface 1cd7a243-c4f2-4cf5-9462-c13718ee7e00; detaching it from the instance and deleting it from the info cache [ 1259.188527] env[61824]: DEBUG nova.network.neutron [req-4d6b7e79-922d-49d0-9e7d-17a8d52c2953 req-eafa435c-8019-47e2-b7e5-5c2b64abbce7 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1259.664479] env[61824]: DEBUG nova.network.neutron [-] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1259.690849] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f88b680-5a9f-4a2e-be07-d0e48398acfc {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.700337] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7bd206-c272-47aa-876f-f2a8a99f3d53 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.721811] env[61824]: DEBUG nova.compute.manager [req-4d6b7e79-922d-49d0-9e7d-17a8d52c2953 req-eafa435c-8019-47e2-b7e5-5c2b64abbce7 service nova] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Detach interface failed, port_id=1cd7a243-c4f2-4cf5-9462-c13718ee7e00, reason: Instance 0baaad18-8cbe-4b2d-872b-ee259534b1fb could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1260.167445] env[61824]: INFO nova.compute.manager [-] [instance: 0baaad18-8cbe-4b2d-872b-ee259534b1fb] Took 1.20 seconds to deallocate network for instance. [ 1260.673603] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1260.673877] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1260.674121] env[61824]: DEBUG nova.objects.instance [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'resources' on Instance uuid 0baaad18-8cbe-4b2d-872b-ee259534b1fb {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1261.217709] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cbee84-657a-4c80-a515-5efe9e1ec891 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.224822] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832be9d3-e4f9-4e95-b854-fbec0b35da7c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.253326] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c5997e-e199-4808-8f1a-1f97070e931e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.260196] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b10f0f-c53a-4b6e-ada2-49277cd5d396 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.274139] env[61824]: DEBUG nova.compute.provider_tree [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1261.777379] env[61824]: DEBUG nova.scheduler.client.report [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1262.282588] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.608s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.300856] env[61824]: INFO nova.scheduler.client.report [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted allocations for instance 0baaad18-8cbe-4b2d-872b-ee259534b1fb [ 1262.808252] env[61824]: DEBUG oslo_concurrency.lockutils [None req-9b8b9f26-588d-4c93-bce8-08b473c60748 tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "0baaad18-8cbe-4b2d-872b-ee259534b1fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.948s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.329346] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.329725] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1266.329725] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1266.432789] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.433063] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.433325] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "8aa00861-bca6-4861-a5e5-0c538155c9ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.433528] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.433700] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1266.435641] env[61824]: INFO nova.compute.manager [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Terminating instance [ 1266.437329] env[61824]: DEBUG nova.compute.manager [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1266.437529] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1266.438387] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a41413e-730e-4451-a8d3-02b43d5dce55 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.446442] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1266.446662] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6dc422d-989c-4dfa-b3de-236fa67f395a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.452743] env[61824]: DEBUG oslo_vmware.api [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1266.452743] env[61824]: value = "task-1276253" [ 1266.452743] env[61824]: _type = "Task" [ 1266.452743] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.459936] env[61824]: DEBUG oslo_vmware.api [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276253, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.963131] env[61824]: DEBUG oslo_vmware.api [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276253, 'name': PowerOffVM_Task, 'duration_secs': 0.178661} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.963470] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1266.963654] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1266.963903] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b04f0881-0d3f-4300-814c-3f86a161565e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.023113] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1267.023399] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1267.023623] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleting the datastore file [datastore2] 8aa00861-bca6-4861-a5e5-0c538155c9ab {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1267.023917] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7cd218b8-1fab-49ef-8b2f-98322978f449 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.030610] env[61824]: DEBUG oslo_vmware.api [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for the task: (returnval){ [ 1267.030610] env[61824]: value = "task-1276255" [ 1267.030610] env[61824]: _type = "Task" [ 1267.030610] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.037745] env[61824]: DEBUG oslo_vmware.api [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.335172] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Didn't find any instances for network info cache update. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1267.335511] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.335559] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.335696] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.335847] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.335988] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.336150] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1267.336277] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1267.336416] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1268.194346] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.194551] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.194703] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.194852] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61824) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1268.195453] env[61824]: DEBUG oslo_vmware.api [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Task: {'id': task-1276255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127146} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.196203] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2440d4a-e022-47a1-b81b-ae9359a1f48b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.198703] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1268.198893] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1268.199082] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1268.199260] env[61824]: INFO nova.compute.manager [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Took 1.76 seconds to destroy the instance on the hypervisor. [ 1268.199488] env[61824]: DEBUG oslo.service.loopingcall [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1268.199672] env[61824]: DEBUG nova.compute.manager [-] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1268.199762] env[61824]: DEBUG nova.network.neutron [-] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1268.206622] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a99f62f-5f90-44ae-947b-c7e95f00766b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.219591] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214554d1-c681-4a06-8182-ffe957902d85 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.225624] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3ad1f5-5bdf-48ee-b728-1e4d7d30b3a3 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.255840] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181089MB free_disk=142GB free_vcpus=48 pci_devices=None {{(pid=61824) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1268.256085] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1268.256245] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1268.478982] env[61824]: DEBUG nova.compute.manager [req-8f1e897c-ef04-4884-9a2e-cd3ddc604bcd req-422bb031-dc80-4188-b909-db345d656ee5 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Received event network-vif-deleted-87ec4b1c-0d02-46ba-878d-b392ae74f472 {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1268.479232] env[61824]: INFO nova.compute.manager [req-8f1e897c-ef04-4884-9a2e-cd3ddc604bcd req-422bb031-dc80-4188-b909-db345d656ee5 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Neutron deleted interface 87ec4b1c-0d02-46ba-878d-b392ae74f472; detaching it from the instance and deleting it from the info cache [ 1268.479411] env[61824]: DEBUG nova.network.neutron [req-8f1e897c-ef04-4884-9a2e-cd3ddc604bcd req-422bb031-dc80-4188-b909-db345d656ee5 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.955711] env[61824]: DEBUG nova.network.neutron [-] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1268.982457] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f6b57ad-93ff-468a-a82f-a72c058d54a6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.992371] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcec4e65-61e4-4f3e-8959-f61474887ba8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.015952] env[61824]: DEBUG nova.compute.manager [req-8f1e897c-ef04-4884-9a2e-cd3ddc604bcd req-422bb031-dc80-4188-b909-db345d656ee5 service nova] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Detach interface failed, port_id=87ec4b1c-0d02-46ba-878d-b392ae74f472, reason: Instance 8aa00861-bca6-4861-a5e5-0c538155c9ab could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1269.278260] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Instance 8aa00861-bca6-4861-a5e5-0c538155c9ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61824) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1269.278472] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1269.278617] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61824) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1269.303882] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec37c920-2324-4d72-82d5-ff57e4097cd0 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.312031] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37dc6937-231d-4386-89f7-68ae9095798f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.341794] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15859e87-2b58-4169-b05a-b6d4ef45daa8 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.348938] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfbf612-e391-4504-be8c-69ea6afec62b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.361915] env[61824]: DEBUG nova.compute.provider_tree [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1269.458017] env[61824]: INFO nova.compute.manager [-] [instance: 8aa00861-bca6-4861-a5e5-0c538155c9ab] Took 1.26 seconds to deallocate network for instance. [ 1269.864775] env[61824]: DEBUG nova.scheduler.client.report [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1269.964439] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1270.370207] env[61824]: DEBUG nova.compute.resource_tracker [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61824) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1270.370500] env[61824]: DEBUG oslo_concurrency.lockutils [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.114s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1270.370744] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.407s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1270.370981] env[61824]: DEBUG nova.objects.instance [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lazy-loading 'resources' on Instance uuid 8aa00861-bca6-4861-a5e5-0c538155c9ab {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1270.901075] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052ad468-47fe-4992-94da-84ae9ac82578 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.909228] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253c7300-1805-426a-b371-cd73b40a2c36 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.938300] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af9d9c81-5c7d-429c-b7e3-b5eb70f1a42c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.945467] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a58286-182d-45a4-b9b9-1b1d607a0849 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.958203] env[61824]: DEBUG nova.compute.provider_tree [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1271.461542] env[61824]: DEBUG nova.scheduler.client.report [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1271.966893] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.596s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1271.983853] env[61824]: INFO nova.scheduler.client.report [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Deleted allocations for instance 8aa00861-bca6-4861-a5e5-0c538155c9ab [ 1272.493027] env[61824]: DEBUG oslo_concurrency.lockutils [None req-8f49c17d-37b5-4943-83b6-bb58ff61983a tempest-ServerActionsTestOtherA-393912163 tempest-ServerActionsTestOtherA-393912163-project-member] Lock "8aa00861-bca6-4861-a5e5-0c538155c9ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.060s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1277.389853] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1277.390183] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1277.893154] env[61824]: DEBUG nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Starting instance... {{(pid=61824) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1278.413817] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.414102] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.415825] env[61824]: INFO nova.compute.claims [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1279.451803] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b3a0c2-223c-4ff2-805b-c97e2ed98217 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.459172] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b4cb09-b7a3-4295-a0cc-d430217e502c {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.488935] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71612ee-0c89-4697-9873-50ec3a19f303 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.495922] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a8ca704-e5e8-4402-beb1-92c49df64687 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.508384] env[61824]: DEBUG nova.compute.provider_tree [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1280.011068] env[61824]: DEBUG nova.scheduler.client.report [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1280.515748] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.101s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.516315] env[61824]: DEBUG nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Start building networks asynchronously for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1281.021496] env[61824]: DEBUG nova.compute.utils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Using /dev/sd instead of None {{(pid=61824) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1281.022976] env[61824]: DEBUG nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Allocating IP information in the background. {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1281.023161] env[61824]: DEBUG nova.network.neutron [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] allocate_for_instance() {{(pid=61824) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1281.070071] env[61824]: DEBUG nova.policy [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a04370540ec42e28dc2d46239e81435', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7bbee86aa2624009a3ff52d0d2decf9b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61824) authorize /opt/stack/nova/nova/policy.py:201}} [ 1281.300667] env[61824]: DEBUG nova.network.neutron [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Successfully created port: b8e57591-d26c-4523-8fe0-64b09823adba {{(pid=61824) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1281.526545] env[61824]: DEBUG nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Start building block device mappings for instance. {{(pid=61824) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1282.536281] env[61824]: DEBUG nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Start spawning the instance on the hypervisor. {{(pid=61824) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1282.561134] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-12T10:51:10Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-12T10:50:53Z,direct_url=,disk_format='vmdk',id=11b2670a-52f3-4a0e-8363-54ca7a7f5264,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='c39dca4dc3bb41789ac7f6f806138c14',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-12T10:50:53Z,virtual_size=,visibility=), allow threads: False {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1282.561388] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Flavor limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1282.561553] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Image limits 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1282.561739] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Flavor pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1282.561888] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Image pref 0:0:0 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1282.562051] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61824) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1282.562275] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1282.562475] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1282.562657] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Got 1 possible topologies {{(pid=61824) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1282.562828] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1282.563013] env[61824]: DEBUG nova.virt.hardware [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61824) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1282.563926] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb40c05-a143-4d83-8736-0a8e4edc47d6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.571876] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f3e696-d19c-4bbf-8ea7-c4869d6dd5ac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.663676] env[61824]: DEBUG nova.compute.manager [req-ad2761ae-1c67-47d5-aaa9-66fab9eae467 req-e00db592-59b5-4817-a60d-deb3e91dd8c6 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Received event network-vif-plugged-b8e57591-d26c-4523-8fe0-64b09823adba {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1282.663897] env[61824]: DEBUG oslo_concurrency.lockutils [req-ad2761ae-1c67-47d5-aaa9-66fab9eae467 req-e00db592-59b5-4817-a60d-deb3e91dd8c6 service nova] Acquiring lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1282.664124] env[61824]: DEBUG oslo_concurrency.lockutils [req-ad2761ae-1c67-47d5-aaa9-66fab9eae467 req-e00db592-59b5-4817-a60d-deb3e91dd8c6 service nova] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1282.664297] env[61824]: DEBUG oslo_concurrency.lockutils [req-ad2761ae-1c67-47d5-aaa9-66fab9eae467 req-e00db592-59b5-4817-a60d-deb3e91dd8c6 service nova] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1282.664467] env[61824]: DEBUG nova.compute.manager [req-ad2761ae-1c67-47d5-aaa9-66fab9eae467 req-e00db592-59b5-4817-a60d-deb3e91dd8c6 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] No waiting events found dispatching network-vif-plugged-b8e57591-d26c-4523-8fe0-64b09823adba {{(pid=61824) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1282.664635] env[61824]: WARNING nova.compute.manager [req-ad2761ae-1c67-47d5-aaa9-66fab9eae467 req-e00db592-59b5-4817-a60d-deb3e91dd8c6 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Received unexpected event network-vif-plugged-b8e57591-d26c-4523-8fe0-64b09823adba for instance with vm_state building and task_state spawning. [ 1283.191145] env[61824]: DEBUG nova.network.neutron [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Successfully updated port: b8e57591-d26c-4523-8fe0-64b09823adba {{(pid=61824) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1283.214160] env[61824]: DEBUG nova.compute.manager [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Received event network-changed-b8e57591-d26c-4523-8fe0-64b09823adba {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1283.214466] env[61824]: DEBUG nova.compute.manager [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Refreshing instance network info cache due to event network-changed-b8e57591-d26c-4523-8fe0-64b09823adba. {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1283.214727] env[61824]: DEBUG oslo_concurrency.lockutils [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] Acquiring lock "refresh_cache-bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.214909] env[61824]: DEBUG oslo_concurrency.lockutils [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] Acquired lock "refresh_cache-bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.215251] env[61824]: DEBUG nova.network.neutron [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Refreshing network info cache for port b8e57591-d26c-4523-8fe0-64b09823adba {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1283.693684] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "refresh_cache-bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1283.745501] env[61824]: DEBUG nova.network.neutron [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1283.810836] env[61824]: DEBUG nova.network.neutron [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.313652] env[61824]: DEBUG oslo_concurrency.lockutils [req-2a204890-5269-4d71-b396-ab895cb38645 req-f0f681a2-24b2-430c-bec9-0f4f2cc23122 service nova] Releasing lock "refresh_cache-bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1284.314032] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquired lock "refresh_cache-bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1284.314391] env[61824]: DEBUG nova.network.neutron [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Building network info cache for instance {{(pid=61824) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1284.845694] env[61824]: DEBUG nova.network.neutron [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Instance cache missing network info. {{(pid=61824) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1284.962545] env[61824]: DEBUG nova.network.neutron [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Updating instance_info_cache with network_info: [{"id": "b8e57591-d26c-4523-8fe0-64b09823adba", "address": "fa:16:3e:f5:7e:6a", "network": {"id": "a9e01ca5-623e-495c-b8b6-ccb285935d14", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1098695332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bbee86aa2624009a3ff52d0d2decf9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60e7ee7b-4d02-4d68-af2e-5ab7d9708120", "external-id": "nsx-vlan-transportzone-550", "segmentation_id": 550, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e57591-d2", "ovs_interfaceid": "b8e57591-d26c-4523-8fe0-64b09823adba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1285.464989] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Releasing lock "refresh_cache-bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1285.465344] env[61824]: DEBUG nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Instance network_info: |[{"id": "b8e57591-d26c-4523-8fe0-64b09823adba", "address": "fa:16:3e:f5:7e:6a", "network": {"id": "a9e01ca5-623e-495c-b8b6-ccb285935d14", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1098695332-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7bbee86aa2624009a3ff52d0d2decf9b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60e7ee7b-4d02-4d68-af2e-5ab7d9708120", "external-id": "nsx-vlan-transportzone-550", "segmentation_id": 550, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8e57591-d2", "ovs_interfaceid": "b8e57591-d26c-4523-8fe0-64b09823adba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61824) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1285.465770] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:7e:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60e7ee7b-4d02-4d68-af2e-5ab7d9708120', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8e57591-d26c-4523-8fe0-64b09823adba', 'vif_model': 'vmxnet3'}] {{(pid=61824) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1285.473190] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Creating folder: Project (7bbee86aa2624009a3ff52d0d2decf9b). Parent ref: group-v274074. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1285.473465] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4726a5e1-cd1b-417a-a338-32218d896bac {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.485535] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Created folder: Project (7bbee86aa2624009a3ff52d0d2decf9b) in parent group-v274074. [ 1285.485719] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Creating folder: Instances. Parent ref: group-v274255. {{(pid=61824) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1285.485938] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-310f2fad-0d3d-45a7-b66b-55dd419d8d20 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.494660] env[61824]: INFO nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Created folder: Instances in parent group-v274255. [ 1285.494872] env[61824]: DEBUG oslo.service.loopingcall [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1285.495071] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Creating VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1285.495262] env[61824]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7a10766-8fa2-410c-931b-6d4e723d4162 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.512910] env[61824]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1285.512910] env[61824]: value = "task-1276258" [ 1285.512910] env[61824]: _type = "Task" [ 1285.512910] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1285.519899] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276258, 'name': CreateVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.022871] env[61824]: DEBUG oslo_vmware.api [-] Task: {'id': task-1276258, 'name': CreateVM_Task, 'duration_secs': 0.274026} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.023313] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Created VM on the ESX host {{(pid=61824) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1286.023729] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.023902] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.024257] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1286.024510] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72286de7-bd99-465b-b1c0-9a6e91c5cb36 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.028659] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1286.028659] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52b81251-638a-d6d7-31a2-1b2cfc475320" [ 1286.028659] env[61824]: _type = "Task" [ 1286.028659] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.035886] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b81251-638a-d6d7-31a2-1b2cfc475320, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1286.539308] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b81251-638a-d6d7-31a2-1b2cfc475320, 'name': SearchDatastore_Task, 'duration_secs': 0.011332} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1286.539486] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1286.539724] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Processing image 11b2670a-52f3-4a0e-8363-54ca7a7f5264 {{(pid=61824) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1286.539955] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.540124] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.540307] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1286.540558] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34f53206-f973-4a92-9160-6af103cac748 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.548335] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61824) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1286.548512] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61824) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1286.549215] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9cee412-3143-4c24-bbed-0f5cc29219eb {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1286.553765] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1286.553765] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]52b9b6f1-3aa8-172d-a755-4b3a439c87c5" [ 1286.553765] env[61824]: _type = "Task" [ 1286.553765] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1286.560602] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b9b6f1-3aa8-172d-a755-4b3a439c87c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.065281] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]52b9b6f1-3aa8-172d-a755-4b3a439c87c5, 'name': SearchDatastore_Task, 'duration_secs': 0.007187} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.066066] env[61824]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f688603b-1405-411d-acfa-a371b0a34f4f {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.070701] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1287.070701] env[61824]: value = "session[52949634-3ad9-62be-d15a-1830fc048cc1]529a64f4-4e6c-c2cf-4fe9-55f5bfe1019d" [ 1287.070701] env[61824]: _type = "Task" [ 1287.070701] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.077612] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529a64f4-4e6c-c2cf-4fe9-55f5bfe1019d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.581304] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': session[52949634-3ad9-62be-d15a-1830fc048cc1]529a64f4-4e6c-c2cf-4fe9-55f5bfe1019d, 'name': SearchDatastore_Task, 'duration_secs': 0.009017} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1287.581546] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk" {{(pid=61824) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.581798] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] bf7bdb24-0811-4818-8e8e-a3df6b4e09fe/bf7bdb24-0811-4818-8e8e-a3df6b4e09fe.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1287.582061] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-739c8712-4698-4e59-a4b1-de510e09cfb6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.588887] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1287.588887] env[61824]: value = "task-1276259" [ 1287.588887] env[61824]: _type = "Task" [ 1287.588887] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.596217] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.098648] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276259, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.411328} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.099025] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/11b2670a-52f3-4a0e-8363-54ca7a7f5264/11b2670a-52f3-4a0e-8363-54ca7a7f5264.vmdk to [datastore2] bf7bdb24-0811-4818-8e8e-a3df6b4e09fe/bf7bdb24-0811-4818-8e8e-a3df6b4e09fe.vmdk {{(pid=61824) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1288.099132] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Extending root virtual disk to 1048576 {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1288.099389] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5136909-fbea-4fb9-8797-df5041167180 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.105575] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1288.105575] env[61824]: value = "task-1276260" [ 1288.105575] env[61824]: _type = "Task" [ 1288.105575] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.112795] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276260, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.615394] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276260, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.139292} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.615662] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Extended root virtual disk {{(pid=61824) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1288.616414] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fe982d-e15d-4259-a86d-94e325d8aaa2 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.637490] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] bf7bdb24-0811-4818-8e8e-a3df6b4e09fe/bf7bdb24-0811-4818-8e8e-a3df6b4e09fe.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1288.637706] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07a4a129-bec0-4fac-ba58-03b4222d13a4 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.656470] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1288.656470] env[61824]: value = "task-1276261" [ 1288.656470] env[61824]: _type = "Task" [ 1288.656470] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.663293] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276261, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.166671] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276261, 'name': ReconfigVM_Task, 'duration_secs': 0.285092} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.167068] env[61824]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Reconfigured VM instance instance-00000070 to attach disk [datastore2] bf7bdb24-0811-4818-8e8e-a3df6b4e09fe/bf7bdb24-0811-4818-8e8e-a3df6b4e09fe.vmdk or device None with type sparse {{(pid=61824) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1289.167611] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5118998b-ec92-448f-bf49-6ed88f45d63e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.174132] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1289.174132] env[61824]: value = "task-1276262" [ 1289.174132] env[61824]: _type = "Task" [ 1289.174132] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.181477] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276262, 'name': Rename_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.683314] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276262, 'name': Rename_Task, 'duration_secs': 0.130777} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.683603] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Powering on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1289.683847] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cd5709a1-7c83-424b-9203-a7a5ad87c5d7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.689770] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1289.689770] env[61824]: value = "task-1276263" [ 1289.689770] env[61824]: _type = "Task" [ 1289.689770] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.696741] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276263, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.199430] env[61824]: DEBUG oslo_vmware.api [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276263, 'name': PowerOnVM_Task, 'duration_secs': 0.431167} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.199874] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Powered on the VM {{(pid=61824) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1290.199942] env[61824]: INFO nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Took 7.66 seconds to spawn the instance on the hypervisor. [ 1290.200089] env[61824]: DEBUG nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Checking state {{(pid=61824) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1290.200903] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7af0604-f08c-499d-ab23-75e4b28a00a9 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.720819] env[61824]: INFO nova.compute.manager [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Took 12.32 seconds to build instance. [ 1291.222613] env[61824]: DEBUG oslo_concurrency.lockutils [None req-c6be40bf-1b95-4ff0-85cb-99b1ba81c1f2 tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.832s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.661062] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1291.661062] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1291.661062] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1291.661062] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1291.661062] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1291.663295] env[61824]: INFO nova.compute.manager [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Terminating instance [ 1291.665045] env[61824]: DEBUG nova.compute.manager [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Start destroying the instance on the hypervisor. {{(pid=61824) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1291.665250] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Destroying instance {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1291.666079] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca283be2-2967-4493-82d4-681c68905bf6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.673746] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Powering off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1291.673976] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b908471f-c755-41da-9727-1303a867555e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.680348] env[61824]: DEBUG oslo_vmware.api [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1291.680348] env[61824]: value = "task-1276264" [ 1291.680348] env[61824]: _type = "Task" [ 1291.680348] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.687944] env[61824]: DEBUG oslo_vmware.api [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276264, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.190997] env[61824]: DEBUG oslo_vmware.api [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276264, 'name': PowerOffVM_Task, 'duration_secs': 0.200436} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.191403] env[61824]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Powered off the VM {{(pid=61824) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1292.191585] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Unregistering the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1292.191840] env[61824]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8e792bc-0462-420c-a4c1-e420b4c3be31 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.255640] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Unregistered the VM {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1292.255994] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Deleting contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1292.256074] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Deleting the datastore file [datastore2] bf7bdb24-0811-4818-8e8e-a3df6b4e09fe {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1292.256328] env[61824]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-673191a9-078e-40ca-ac79-724653cbfdd7 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.263564] env[61824]: DEBUG oslo_vmware.api [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for the task: (returnval){ [ 1292.263564] env[61824]: value = "task-1276266" [ 1292.263564] env[61824]: _type = "Task" [ 1292.263564] env[61824]: } to complete. {{(pid=61824) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.270709] env[61824]: DEBUG oslo_vmware.api [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.773578] env[61824]: DEBUG oslo_vmware.api [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Task: {'id': task-1276266, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137825} completed successfully. {{(pid=61824) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.773578] env[61824]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Deleted the datastore file {{(pid=61824) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1292.773798] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Deleted contents of the VM from datastore datastore2 {{(pid=61824) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1292.773999] env[61824]: DEBUG nova.virt.vmwareapi.vmops [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Instance destroyed {{(pid=61824) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1292.774108] env[61824]: INFO nova.compute.manager [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1292.774294] env[61824]: DEBUG oslo.service.loopingcall [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61824) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1292.774487] env[61824]: DEBUG nova.compute.manager [-] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Deallocating network for instance {{(pid=61824) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1292.774583] env[61824]: DEBUG nova.network.neutron [-] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] deallocate_for_instance() {{(pid=61824) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1293.023239] env[61824]: DEBUG nova.compute.manager [req-fbc4b8a7-36ff-4375-aa9a-dba0e65ce183 req-0d0592cc-2085-4c39-8297-5f3fe70df884 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Received event network-vif-deleted-b8e57591-d26c-4523-8fe0-64b09823adba {{(pid=61824) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1293.023454] env[61824]: INFO nova.compute.manager [req-fbc4b8a7-36ff-4375-aa9a-dba0e65ce183 req-0d0592cc-2085-4c39-8297-5f3fe70df884 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Neutron deleted interface b8e57591-d26c-4523-8fe0-64b09823adba; detaching it from the instance and deleting it from the info cache [ 1293.023635] env[61824]: DEBUG nova.network.neutron [req-fbc4b8a7-36ff-4375-aa9a-dba0e65ce183 req-0d0592cc-2085-4c39-8297-5f3fe70df884 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1293.499735] env[61824]: DEBUG nova.network.neutron [-] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Updating instance_info_cache with network_info: [] {{(pid=61824) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1293.526485] env[61824]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5c570c1-9594-47bb-a798-78a0e08a6dea {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.536303] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b769ed-55ad-4c6d-8066-b29910aac3c6 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.558206] env[61824]: DEBUG nova.compute.manager [req-fbc4b8a7-36ff-4375-aa9a-dba0e65ce183 req-0d0592cc-2085-4c39-8297-5f3fe70df884 service nova] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Detach interface failed, port_id=b8e57591-d26c-4523-8fe0-64b09823adba, reason: Instance bf7bdb24-0811-4818-8e8e-a3df6b4e09fe could not be found. {{(pid=61824) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1294.002784] env[61824]: INFO nova.compute.manager [-] [instance: bf7bdb24-0811-4818-8e8e-a3df6b4e09fe] Took 1.23 seconds to deallocate network for instance. [ 1294.509910] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1294.510225] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1294.510467] env[61824]: DEBUG nova.objects.instance [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lazy-loading 'resources' on Instance uuid bf7bdb24-0811-4818-8e8e-a3df6b4e09fe {{(pid=61824) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1295.042824] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3ca1af-7aee-4bc8-9dcd-624e206f4f8a {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.050331] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06d489c-56a8-425d-a27f-44c023f4a13b {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.079318] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021a00a8-c20e-4b67-ba02-30ae77fd42d5 {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.086316] env[61824]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cea6c50-71af-40b2-82a5-0d669eae8d1e {{(pid=61824) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.098919] env[61824]: DEBUG nova.compute.provider_tree [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Inventory has not changed in ProviderTree for provider: 8d9ece0e-729c-4d69-b741-d89969200860 {{(pid=61824) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1295.602407] env[61824]: DEBUG nova.scheduler.client.report [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Inventory has not changed for provider 8d9ece0e-729c-4d69-b741-d89969200860 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61824) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1296.107651] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.597s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1296.124889] env[61824]: INFO nova.scheduler.client.report [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Deleted allocations for instance bf7bdb24-0811-4818-8e8e-a3df6b4e09fe [ 1296.631755] env[61824]: DEBUG oslo_concurrency.lockutils [None req-b0e34b5c-5930-4a89-b86f-2486406e15ff tempest-ServerPasswordTestJSON-512224981 tempest-ServerPasswordTestJSON-512224981-project-member] Lock "bf7bdb24-0811-4818-8e8e-a3df6b4e09fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.971s {{(pid=61824) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1321.108197] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.616386] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.616646] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Starting heal instance info cache {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1321.616646] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Rebuilding the list of instances to heal {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1322.121922] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Didn't find any instances for network info cache update. {{(pid=61824) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1322.122260] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.122358] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.124343] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.124343] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.124343] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.124343] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1322.124343] env[61824]: DEBUG nova.compute.manager [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61824) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1322.124343] env[61824]: DEBUG oslo_service.periodic_task [None req-1046ac3d-1275-4a07-a59b-0b7fb84f5f9e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61824) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}